I0424 23:37:17.679805 8 test_context.go:423] Tolerating taints "node-role.kubernetes.io/master" when considering if nodes are ready I0424 23:37:17.679990 8 e2e.go:124] Starting e2e run "89e93e23-fc35-4cb8-a25e-45b45f06c15b" on Ginkgo node 1 {"msg":"Test Suite starting","total":275,"completed":0,"skipped":0,"failed":0} Running Suite: Kubernetes e2e suite =================================== Random Seed: 1587771436 - Will randomize all specs Will run 275 of 4992 specs Apr 24 23:37:17.741: INFO: >>> kubeConfig: /root/.kube/config Apr 24 23:37:17.746: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable Apr 24 23:37:17.773: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready Apr 24 23:37:17.806: INFO: 12 / 12 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) Apr 24 23:37:17.806: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. Apr 24 23:37:17.806: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start Apr 24 23:37:17.825: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kindnet' (0 seconds elapsed) Apr 24 23:37:17.825: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) Apr 24 23:37:17.825: INFO: e2e test version: v1.19.0-alpha.0.779+84dc7046797aad Apr 24 23:37:17.826: INFO: kube-apiserver version: v1.17.0 Apr 24 23:37:17.826: INFO: >>> kubeConfig: /root/.kube/config Apr 24 23:37:17.829: INFO: Cluster IP family: ipv4 SSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:37:17.830: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container Apr 24 23:37:17.894: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled. STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153 [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod Apr 24 23:37:17.896: INFO: PodSpec: initContainers in spec.initContainers Apr 24 23:38:08.343: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-ecb3f337-aaa1-42e0-9c40-c307226c2e43", GenerateName:"", Namespace:"init-container-4454", SelfLink:"/api/v1/namespaces/init-container-4454/pods/pod-init-ecb3f337-aaa1-42e0-9c40-c307226c2e43", UID:"56a8a004-8a58-4e10-a5be-201b93da286c", ResourceVersion:"10776693", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63723368237, loc:(*time.Location)(0x7b1e080)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"896318625"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-z9p2f", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc001e4e5c0), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-z9p2f", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-z9p2f", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.2", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-z9p2f", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc0022cd488), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"latest-worker", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc001d42e70), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0022cd510)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0022cd530)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0022cd538), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0022cd53c), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368237, loc:(*time.Location)(0x7b1e080)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368237, loc:(*time.Location)(0x7b1e080)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368237, loc:(*time.Location)(0x7b1e080)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368237, loc:(*time.Location)(0x7b1e080)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"172.17.0.13", PodIP:"10.244.2.48", PodIPs:[]v1.PodIP{v1.PodIP{IP:"10.244.2.48"}}, StartTime:(*v1.Time)(0xc002ba61a0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc001d42f50)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://cd5b727d9492f614dabe650a9042a84950d121b94a5ac57ba0fd6860089d6593", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc002ba61e0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc002ba61c0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.2", ImageID:"", ContainerID:"", Started:(*bool)(0xc0022cd5bf)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}} [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:08.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-4454" for this suite. • [SLOW TEST:50.524 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","total":275,"completed":1,"skipped":9,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:08.354: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 24 23:38:08.427: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config version' Apr 24 23:38:08.571: INFO: stderr: "" Apr 24 23:38:08.571: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"19+\", GitVersion:\"v1.19.0-alpha.0.779+84dc7046797aad\", GitCommit:\"84dc7046797aad80f258b6740a98e79199c8bb4d\", GitTreeState:\"clean\", BuildDate:\"2020-03-15T16:56:42Z\", GoVersion:\"go1.13.8\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"17\", GitVersion:\"v1.17.0\", GitCommit:\"70132b0f130acc0bed193d9ba59dd186f0e634cf\", GitTreeState:\"clean\", BuildDate:\"2020-01-14T00:09:19Z\", GoVersion:\"go1.13.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:08.571: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6610" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance]","total":275,"completed":2,"skipped":71,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should support configurable pod DNS nameservers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:08.580: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should support configurable pod DNS nameservers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod with dnsPolicy=None and customized dnsConfig... Apr 24 23:38:08.698: INFO: Created pod &Pod{ObjectMeta:{dns-5306 dns-5306 /api/v1/namespaces/dns-5306/pods/dns-5306 23250451-094f-4c73-ad8b-59c94ae4f8ff 10776704 0 2020-04-24 23:38:08 +0000 UTC map[] map[] [] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-7mf8s,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-7mf8s,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-7mf8s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 24 23:38:08.718: INFO: The status of Pod dns-5306 is Pending, waiting for it to be Running (with Ready = true) Apr 24 23:38:10.722: INFO: The status of Pod dns-5306 is Pending, waiting for it to be Running (with Ready = true) Apr 24 23:38:12.729: INFO: The status of Pod dns-5306 is Running (Ready = true) STEP: Verifying customized DNS suffix list is configured on pod... Apr 24 23:38:12.729: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-5306 PodName:dns-5306 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 24 23:38:12.729: INFO: >>> kubeConfig: /root/.kube/config I0424 23:38:12.757813 8 log.go:172] (0xc002ffc2c0) (0xc001097040) Create stream I0424 23:38:12.757848 8 log.go:172] (0xc002ffc2c0) (0xc001097040) Stream added, broadcasting: 1 I0424 23:38:12.760600 8 log.go:172] (0xc002ffc2c0) Reply frame received for 1 I0424 23:38:12.760635 8 log.go:172] (0xc002ffc2c0) (0xc0011306e0) Create stream I0424 23:38:12.760649 8 log.go:172] (0xc002ffc2c0) (0xc0011306e0) Stream added, broadcasting: 3 I0424 23:38:12.761859 8 log.go:172] (0xc002ffc2c0) Reply frame received for 3 I0424 23:38:12.761894 8 log.go:172] (0xc002ffc2c0) (0xc000d820a0) Create stream I0424 23:38:12.761908 8 log.go:172] (0xc002ffc2c0) (0xc000d820a0) Stream added, broadcasting: 5 I0424 23:38:12.762898 8 log.go:172] (0xc002ffc2c0) Reply frame received for 5 I0424 23:38:12.856624 8 log.go:172] (0xc002ffc2c0) Data frame received for 3 I0424 23:38:12.856655 8 log.go:172] (0xc0011306e0) (3) Data frame handling I0424 23:38:12.856674 8 log.go:172] (0xc0011306e0) (3) Data frame sent I0424 23:38:12.857979 8 log.go:172] (0xc002ffc2c0) Data frame received for 3 I0424 23:38:12.858018 8 log.go:172] (0xc0011306e0) (3) Data frame handling I0424 23:38:12.858042 8 log.go:172] (0xc002ffc2c0) Data frame received for 5 I0424 23:38:12.858053 8 log.go:172] (0xc000d820a0) (5) Data frame handling I0424 23:38:12.860547 8 log.go:172] (0xc002ffc2c0) Data frame received for 1 I0424 23:38:12.860596 8 log.go:172] (0xc001097040) (1) Data frame handling I0424 23:38:12.860618 8 log.go:172] (0xc001097040) (1) Data frame sent I0424 23:38:12.860650 8 log.go:172] (0xc002ffc2c0) (0xc001097040) Stream removed, broadcasting: 1 I0424 23:38:12.860998 8 log.go:172] (0xc002ffc2c0) Go away received I0424 23:38:12.861482 8 log.go:172] (0xc002ffc2c0) (0xc001097040) Stream removed, broadcasting: 1 I0424 23:38:12.861505 8 log.go:172] (0xc002ffc2c0) (0xc0011306e0) Stream removed, broadcasting: 3 I0424 23:38:12.861524 8 log.go:172] (0xc002ffc2c0) (0xc000d820a0) Stream removed, broadcasting: 5 STEP: Verifying customized DNS server is configured on pod... Apr 24 23:38:12.861: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-5306 PodName:dns-5306 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 24 23:38:12.861: INFO: >>> kubeConfig: /root/.kube/config I0424 23:38:12.896447 8 log.go:172] (0xc00245f6b0) (0xc001260500) Create stream I0424 23:38:12.896477 8 log.go:172] (0xc00245f6b0) (0xc001260500) Stream added, broadcasting: 1 I0424 23:38:12.899580 8 log.go:172] (0xc00245f6b0) Reply frame received for 1 I0424 23:38:12.899631 8 log.go:172] (0xc00245f6b0) (0xc0011cc140) Create stream I0424 23:38:12.899646 8 log.go:172] (0xc00245f6b0) (0xc0011cc140) Stream added, broadcasting: 3 I0424 23:38:12.900704 8 log.go:172] (0xc00245f6b0) Reply frame received for 3 I0424 23:38:12.900737 8 log.go:172] (0xc00245f6b0) (0xc0011cc460) Create stream I0424 23:38:12.900746 8 log.go:172] (0xc00245f6b0) (0xc0011cc460) Stream added, broadcasting: 5 I0424 23:38:12.901881 8 log.go:172] (0xc00245f6b0) Reply frame received for 5 I0424 23:38:12.963709 8 log.go:172] (0xc00245f6b0) Data frame received for 3 I0424 23:38:12.963745 8 log.go:172] (0xc0011cc140) (3) Data frame handling I0424 23:38:12.963770 8 log.go:172] (0xc0011cc140) (3) Data frame sent I0424 23:38:12.964967 8 log.go:172] (0xc00245f6b0) Data frame received for 5 I0424 23:38:12.964992 8 log.go:172] (0xc0011cc460) (5) Data frame handling I0424 23:38:12.965036 8 log.go:172] (0xc00245f6b0) Data frame received for 3 I0424 23:38:12.965057 8 log.go:172] (0xc0011cc140) (3) Data frame handling I0424 23:38:12.966216 8 log.go:172] (0xc00245f6b0) Data frame received for 1 I0424 23:38:12.966240 8 log.go:172] (0xc001260500) (1) Data frame handling I0424 23:38:12.966260 8 log.go:172] (0xc001260500) (1) Data frame sent I0424 23:38:12.966282 8 log.go:172] (0xc00245f6b0) (0xc001260500) Stream removed, broadcasting: 1 I0424 23:38:12.966338 8 log.go:172] (0xc00245f6b0) (0xc001260500) Stream removed, broadcasting: 1 I0424 23:38:12.966351 8 log.go:172] (0xc00245f6b0) (0xc0011cc140) Stream removed, broadcasting: 3 I0424 23:38:12.966364 8 log.go:172] (0xc00245f6b0) (0xc0011cc460) Stream removed, broadcasting: 5 Apr 24 23:38:12.966: INFO: Deleting pod dns-5306... I0424 23:38:12.966736 8 log.go:172] (0xc00245f6b0) Go away received [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:12.976: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-5306" for this suite. •{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":275,"completed":3,"skipped":85,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:13.047: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a service in the namespace STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there is no service in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:19.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-8031" for this suite. STEP: Destroying namespace "nsdeletetest-7393" for this suite. Apr 24 23:38:19.622: INFO: Namespace nsdeletetest-7393 was already deleted STEP: Destroying namespace "nsdeletetest-4092" for this suite. • [SLOW TEST:6.578 seconds] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","total":275,"completed":4,"skipped":114,"failed":0} SSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:19.626: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of different groups [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation Apr 24 23:38:19.714: INFO: >>> kubeConfig: /root/.kube/config Apr 24 23:38:22.657: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:32.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-8990" for this suite. • [SLOW TEST:12.736 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of different groups [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","total":275,"completed":5,"skipped":124,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:32.363: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153 [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod Apr 24 23:38:32.406: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:37.437: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-8756" for this suite. • [SLOW TEST:5.126 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","total":275,"completed":6,"skipped":157,"failed":0} SSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:37.489: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0777 on tmpfs Apr 24 23:38:37.546: INFO: Waiting up to 5m0s for pod "pod-73e51e8f-4e93-40c4-99bd-487362187510" in namespace "emptydir-1528" to be "Succeeded or Failed" Apr 24 23:38:37.572: INFO: Pod "pod-73e51e8f-4e93-40c4-99bd-487362187510": Phase="Pending", Reason="", readiness=false. Elapsed: 25.608348ms Apr 24 23:38:39.576: INFO: Pod "pod-73e51e8f-4e93-40c4-99bd-487362187510": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030097605s Apr 24 23:38:41.581: INFO: Pod "pod-73e51e8f-4e93-40c4-99bd-487362187510": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034710557s STEP: Saw pod success Apr 24 23:38:41.581: INFO: Pod "pod-73e51e8f-4e93-40c4-99bd-487362187510" satisfied condition "Succeeded or Failed" Apr 24 23:38:41.584: INFO: Trying to get logs from node latest-worker pod pod-73e51e8f-4e93-40c4-99bd-487362187510 container test-container: STEP: delete the pod Apr 24 23:38:41.618: INFO: Waiting for pod pod-73e51e8f-4e93-40c4-99bd-487362187510 to disappear Apr 24 23:38:41.622: INFO: Pod pod-73e51e8f-4e93-40c4-99bd-487362187510 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:41.622: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1528" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":7,"skipped":165,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:41.628: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0644 on node default medium Apr 24 23:38:41.714: INFO: Waiting up to 5m0s for pod "pod-6a7dfbac-e44e-4a12-87d6-d121a7951952" in namespace "emptydir-827" to be "Succeeded or Failed" Apr 24 23:38:41.725: INFO: Pod "pod-6a7dfbac-e44e-4a12-87d6-d121a7951952": Phase="Pending", Reason="", readiness=false. Elapsed: 11.152107ms Apr 24 23:38:43.730: INFO: Pod "pod-6a7dfbac-e44e-4a12-87d6-d121a7951952": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015741425s Apr 24 23:38:45.734: INFO: Pod "pod-6a7dfbac-e44e-4a12-87d6-d121a7951952": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020067376s STEP: Saw pod success Apr 24 23:38:45.734: INFO: Pod "pod-6a7dfbac-e44e-4a12-87d6-d121a7951952" satisfied condition "Succeeded or Failed" Apr 24 23:38:45.738: INFO: Trying to get logs from node latest-worker2 pod pod-6a7dfbac-e44e-4a12-87d6-d121a7951952 container test-container: STEP: delete the pod Apr 24 23:38:45.781: INFO: Waiting for pod pod-6a7dfbac-e44e-4a12-87d6-d121a7951952 to disappear Apr 24 23:38:45.792: INFO: Pod pod-6a7dfbac-e44e-4a12-87d6-d121a7951952 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:38:45.793: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-827" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":8,"skipped":176,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:38:45.801: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod pod-subpath-test-configmap-9wlj STEP: Creating a pod to test atomic-volume-subpath Apr 24 23:38:45.890: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-9wlj" in namespace "subpath-4944" to be "Succeeded or Failed" Apr 24 23:38:45.909: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Pending", Reason="", readiness=false. Elapsed: 18.578181ms Apr 24 23:38:47.913: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022901739s Apr 24 23:38:49.917: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 4.026582411s Apr 24 23:38:51.921: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 6.030327209s Apr 24 23:38:53.968: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 8.07731399s Apr 24 23:38:55.972: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 10.081418183s Apr 24 23:38:57.976: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 12.085672551s Apr 24 23:38:59.980: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 14.089894664s Apr 24 23:39:01.984: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 16.093709202s Apr 24 23:39:03.988: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 18.097155072s Apr 24 23:39:05.992: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 20.101412422s Apr 24 23:39:07.996: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Running", Reason="", readiness=true. Elapsed: 22.105629758s Apr 24 23:39:10.000: INFO: Pod "pod-subpath-test-configmap-9wlj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.109240881s STEP: Saw pod success Apr 24 23:39:10.000: INFO: Pod "pod-subpath-test-configmap-9wlj" satisfied condition "Succeeded or Failed" Apr 24 23:39:10.002: INFO: Trying to get logs from node latest-worker pod pod-subpath-test-configmap-9wlj container test-container-subpath-configmap-9wlj: STEP: delete the pod Apr 24 23:39:10.035: INFO: Waiting for pod pod-subpath-test-configmap-9wlj to disappear Apr 24 23:39:10.214: INFO: Pod pod-subpath-test-configmap-9wlj no longer exists STEP: Deleting pod pod-subpath-test-configmap-9wlj Apr 24 23:39:10.214: INFO: Deleting pod "pod-subpath-test-configmap-9wlj" in namespace "subpath-4944" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:10.507: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-4944" for this suite. • [SLOW TEST:24.721 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]","total":275,"completed":9,"skipped":193,"failed":0} SSS ------------------------------ [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:10.522: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to update and delete ResourceQuota. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a ResourceQuota STEP: Getting a ResourceQuota STEP: Updating a ResourceQuota STEP: Verifying a ResourceQuota was modified STEP: Deleting a ResourceQuota STEP: Verifying the deleted ResourceQuota [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:10.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-4549" for this suite. •{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","total":275,"completed":10,"skipped":196,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:10.682: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-ef7eeb55-df4e-45a9-a80d-a67ea5f8afc5 STEP: Creating a pod to test consume secrets Apr 24 23:39:10.744: INFO: Waiting up to 5m0s for pod "pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a" in namespace "secrets-8224" to be "Succeeded or Failed" Apr 24 23:39:10.749: INFO: Pod "pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.710728ms Apr 24 23:39:12.753: INFO: Pod "pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009010672s Apr 24 23:39:15.323: INFO: Pod "pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.579590549s STEP: Saw pod success Apr 24 23:39:15.323: INFO: Pod "pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a" satisfied condition "Succeeded or Failed" Apr 24 23:39:15.327: INFO: Trying to get logs from node latest-worker pod pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a container secret-env-test: STEP: delete the pod Apr 24 23:39:15.455: INFO: Waiting for pod pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a to disappear Apr 24 23:39:15.467: INFO: Pod pod-secrets-2307019e-daa8-44fa-9bde-b494e5471f3a no longer exists [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:15.467: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-8224" for this suite. •{"msg":"PASSED [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","total":275,"completed":11,"skipped":215,"failed":0} ------------------------------ [sig-api-machinery] Secrets should patch a secret [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:15.475: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should patch a secret [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a secret STEP: listing secrets in all namespaces to ensure that there are more than zero STEP: patching the secret STEP: deleting the secret using a LabelSelector STEP: listing secrets in all namespaces, searching for label name and value in patch [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:15.582: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3439" for this suite. •{"msg":"PASSED [sig-api-machinery] Secrets should patch a secret [Conformance]","total":275,"completed":12,"skipped":215,"failed":0} SSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:15.590: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-map-e1992a29-1d19-4625-bcb4-01b0c3489be1 STEP: Creating a pod to test consume secrets Apr 24 23:39:15.649: INFO: Waiting up to 5m0s for pod "pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb" in namespace "secrets-6097" to be "Succeeded or Failed" Apr 24 23:39:15.667: INFO: Pod "pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb": Phase="Pending", Reason="", readiness=false. Elapsed: 18.473863ms Apr 24 23:39:17.671: INFO: Pod "pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022255102s Apr 24 23:39:19.674: INFO: Pod "pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025026672s STEP: Saw pod success Apr 24 23:39:19.674: INFO: Pod "pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb" satisfied condition "Succeeded or Failed" Apr 24 23:39:19.676: INFO: Trying to get logs from node latest-worker2 pod pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb container secret-volume-test: STEP: delete the pod Apr 24 23:39:19.711: INFO: Waiting for pod pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb to disappear Apr 24 23:39:19.716: INFO: Pod pod-secrets-95da45d8-92bf-4eb5-b320-2544c58330fb no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:19.716: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-6097" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":275,"completed":13,"skipped":224,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:19.723: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward api env vars Apr 24 23:39:19.778: INFO: Waiting up to 5m0s for pod "downward-api-c6d4487d-2c85-4311-a692-905fad59eae6" in namespace "downward-api-3325" to be "Succeeded or Failed" Apr 24 23:39:19.793: INFO: Pod "downward-api-c6d4487d-2c85-4311-a692-905fad59eae6": Phase="Pending", Reason="", readiness=false. Elapsed: 15.302024ms Apr 24 23:39:21.819: INFO: Pod "downward-api-c6d4487d-2c85-4311-a692-905fad59eae6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.041039961s Apr 24 23:39:23.823: INFO: Pod "downward-api-c6d4487d-2c85-4311-a692-905fad59eae6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045381122s STEP: Saw pod success Apr 24 23:39:23.823: INFO: Pod "downward-api-c6d4487d-2c85-4311-a692-905fad59eae6" satisfied condition "Succeeded or Failed" Apr 24 23:39:23.826: INFO: Trying to get logs from node latest-worker2 pod downward-api-c6d4487d-2c85-4311-a692-905fad59eae6 container dapi-container: STEP: delete the pod Apr 24 23:39:23.905: INFO: Waiting for pod downward-api-c6d4487d-2c85-4311-a692-905fad59eae6 to disappear Apr 24 23:39:23.907: INFO: Pod downward-api-c6d4487d-2c85-4311-a692-905fad59eae6 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:23.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-3325" for this suite. •{"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","total":275,"completed":14,"skipped":236,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:23.916: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a pod in the namespace STEP: Waiting for the pod to have running status STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there are no pods in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:39.132: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-9639" for this suite. STEP: Destroying namespace "nsdeletetest-1189" for this suite. Apr 24 23:39:39.156: INFO: Namespace nsdeletetest-1189 was already deleted STEP: Destroying namespace "nsdeletetest-7029" for this suite. • [SLOW TEST:15.244 seconds] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","total":275,"completed":15,"skipped":294,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:39.160: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a service. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Service STEP: Ensuring resource quota status captures service creation STEP: Deleting a Service STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:39:50.387: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-7346" for this suite. • [SLOW TEST:11.237 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a service. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","total":275,"completed":16,"skipped":302,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:39:50.398: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6966.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6966.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-6966.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-6966.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6966.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 87.7.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.7.87_udp@PTR;check="$$(dig +tcp +noall +answer +search 87.7.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.7.87_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6966.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6966.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-6966.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6966.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-6966.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6966.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 87.7.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.7.87_udp@PTR;check="$$(dig +tcp +noall +answer +search 87.7.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.7.87_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 24 23:39:56.542: INFO: Unable to read wheezy_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.546: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.549: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.552: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.574: INFO: Unable to read jessie_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.576: INFO: Unable to read jessie_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.579: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.582: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:39:56.619: INFO: Lookups using dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958 failed for: [wheezy_udp@dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_udp@dns-test-service.dns-6966.svc.cluster.local jessie_tcp@dns-test-service.dns-6966.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local] Apr 24 23:40:01.623: INFO: Unable to read wheezy_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.626: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.629: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.632: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.652: INFO: Unable to read jessie_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.655: INFO: Unable to read jessie_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.658: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.661: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:01.680: INFO: Lookups using dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958 failed for: [wheezy_udp@dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_udp@dns-test-service.dns-6966.svc.cluster.local jessie_tcp@dns-test-service.dns-6966.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local] Apr 24 23:40:06.623: INFO: Unable to read wheezy_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.627: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.630: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.633: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.660: INFO: Unable to read jessie_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.662: INFO: Unable to read jessie_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.665: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.667: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:06.685: INFO: Lookups using dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958 failed for: [wheezy_udp@dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_udp@dns-test-service.dns-6966.svc.cluster.local jessie_tcp@dns-test-service.dns-6966.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local] Apr 24 23:40:11.623: INFO: Unable to read wheezy_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.627: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.630: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.633: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.652: INFO: Unable to read jessie_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.655: INFO: Unable to read jessie_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.658: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.660: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:11.677: INFO: Lookups using dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958 failed for: [wheezy_udp@dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_udp@dns-test-service.dns-6966.svc.cluster.local jessie_tcp@dns-test-service.dns-6966.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local] Apr 24 23:40:16.623: INFO: Unable to read wheezy_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.626: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.629: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.633: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.654: INFO: Unable to read jessie_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.656: INFO: Unable to read jessie_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.659: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.662: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:16.678: INFO: Lookups using dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958 failed for: [wheezy_udp@dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_udp@dns-test-service.dns-6966.svc.cluster.local jessie_tcp@dns-test-service.dns-6966.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local] Apr 24 23:40:21.676: INFO: Unable to read wheezy_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.680: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.683: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.686: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.705: INFO: Unable to read jessie_udp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.707: INFO: Unable to read jessie_tcp@dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.710: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.713: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local from pod dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958: the server could not find the requested resource (get pods dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958) Apr 24 23:40:21.729: INFO: Lookups using dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958 failed for: [wheezy_udp@dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@dns-test-service.dns-6966.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_udp@dns-test-service.dns-6966.svc.cluster.local jessie_tcp@dns-test-service.dns-6966.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6966.svc.cluster.local] Apr 24 23:40:26.684: INFO: DNS probes using dns-6966/dns-test-51e80b1d-d435-4e7d-87b8-8f3fd1496958 succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:40:27.352: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-6966" for this suite. • [SLOW TEST:36.972 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for services [Conformance]","total":275,"completed":17,"skipped":318,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:40:27.370: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating the pod Apr 24 23:40:32.001: INFO: Successfully updated pod "labelsupdate7e402403-5f10-44bb-8c1a-11fc9fa8c490" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:40:34.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-341" for this suite. • [SLOW TEST:6.667 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":275,"completed":18,"skipped":330,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:40:34.038: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 24 23:40:34.128: INFO: Waiting up to 5m0s for pod "downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7" in namespace "projected-4852" to be "Succeeded or Failed" Apr 24 23:40:34.138: INFO: Pod "downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7": Phase="Pending", Reason="", readiness=false. Elapsed: 9.861774ms Apr 24 23:40:36.142: INFO: Pod "downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014004093s Apr 24 23:40:38.147: INFO: Pod "downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019074585s STEP: Saw pod success Apr 24 23:40:38.147: INFO: Pod "downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7" satisfied condition "Succeeded or Failed" Apr 24 23:40:38.151: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7 container client-container: STEP: delete the pod Apr 24 23:40:38.232: INFO: Waiting for pod downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7 to disappear Apr 24 23:40:38.235: INFO: Pod downwardapi-volume-47fa41ae-c366-4b15-908a-f6562f73d9b7 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:40:38.235: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4852" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","total":275,"completed":19,"skipped":356,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:40:38.326: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0666 on tmpfs Apr 24 23:40:38.388: INFO: Waiting up to 5m0s for pod "pod-b6dab66b-d981-409d-908d-05c8785862ed" in namespace "emptydir-2415" to be "Succeeded or Failed" Apr 24 23:40:38.403: INFO: Pod "pod-b6dab66b-d981-409d-908d-05c8785862ed": Phase="Pending", Reason="", readiness=false. Elapsed: 14.898423ms Apr 24 23:40:40.408: INFO: Pod "pod-b6dab66b-d981-409d-908d-05c8785862ed": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01970947s Apr 24 23:40:42.412: INFO: Pod "pod-b6dab66b-d981-409d-908d-05c8785862ed": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024091643s STEP: Saw pod success Apr 24 23:40:42.412: INFO: Pod "pod-b6dab66b-d981-409d-908d-05c8785862ed" satisfied condition "Succeeded or Failed" Apr 24 23:40:42.415: INFO: Trying to get logs from node latest-worker pod pod-b6dab66b-d981-409d-908d-05c8785862ed container test-container: STEP: delete the pod Apr 24 23:40:42.450: INFO: Waiting for pod pod-b6dab66b-d981-409d-908d-05c8785862ed to disappear Apr 24 23:40:42.457: INFO: Pod pod-b6dab66b-d981-409d-908d-05c8785862ed no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:40:42.457: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2415" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":20,"skipped":396,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:40:42.466: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating Pod STEP: Waiting for the pod running STEP: Geting the pod STEP: Reading file content from the nginx-container Apr 24 23:40:46.609: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-2821 PodName:pod-sharedvolume-1855d1db-59c0-4bbd-b8ac-c5d02738a919 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 24 23:40:46.609: INFO: >>> kubeConfig: /root/.kube/config I0424 23:40:46.651011 8 log.go:172] (0xc00245fef0) (0xc0019f63c0) Create stream I0424 23:40:46.651048 8 log.go:172] (0xc00245fef0) (0xc0019f63c0) Stream added, broadcasting: 1 I0424 23:40:46.653885 8 log.go:172] (0xc00245fef0) Reply frame received for 1 I0424 23:40:46.653948 8 log.go:172] (0xc00245fef0) (0xc0016b5720) Create stream I0424 23:40:46.653968 8 log.go:172] (0xc00245fef0) (0xc0016b5720) Stream added, broadcasting: 3 I0424 23:40:46.655184 8 log.go:172] (0xc00245fef0) Reply frame received for 3 I0424 23:40:46.655250 8 log.go:172] (0xc00245fef0) (0xc001733720) Create stream I0424 23:40:46.655280 8 log.go:172] (0xc00245fef0) (0xc001733720) Stream added, broadcasting: 5 I0424 23:40:46.656498 8 log.go:172] (0xc00245fef0) Reply frame received for 5 I0424 23:40:46.730568 8 log.go:172] (0xc00245fef0) Data frame received for 5 I0424 23:40:46.730594 8 log.go:172] (0xc001733720) (5) Data frame handling I0424 23:40:46.730636 8 log.go:172] (0xc00245fef0) Data frame received for 3 I0424 23:40:46.730671 8 log.go:172] (0xc0016b5720) (3) Data frame handling I0424 23:40:46.730693 8 log.go:172] (0xc0016b5720) (3) Data frame sent I0424 23:40:46.730705 8 log.go:172] (0xc00245fef0) Data frame received for 3 I0424 23:40:46.730719 8 log.go:172] (0xc0016b5720) (3) Data frame handling I0424 23:40:46.732216 8 log.go:172] (0xc00245fef0) Data frame received for 1 I0424 23:40:46.732238 8 log.go:172] (0xc0019f63c0) (1) Data frame handling I0424 23:40:46.732252 8 log.go:172] (0xc0019f63c0) (1) Data frame sent I0424 23:40:46.732266 8 log.go:172] (0xc00245fef0) (0xc0019f63c0) Stream removed, broadcasting: 1 I0424 23:40:46.732353 8 log.go:172] (0xc00245fef0) Go away received I0424 23:40:46.732418 8 log.go:172] (0xc00245fef0) (0xc0019f63c0) Stream removed, broadcasting: 1 I0424 23:40:46.732446 8 log.go:172] (0xc00245fef0) (0xc0016b5720) Stream removed, broadcasting: 3 I0424 23:40:46.732467 8 log.go:172] (0xc00245fef0) (0xc001733720) Stream removed, broadcasting: 5 Apr 24 23:40:46.732: INFO: Exec stderr: "" [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:40:46.732: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2821" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":275,"completed":21,"skipped":408,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:40:46.742: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0644 on node default medium Apr 24 23:40:46.846: INFO: Waiting up to 5m0s for pod "pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802" in namespace "emptydir-3658" to be "Succeeded or Failed" Apr 24 23:40:46.874: INFO: Pod "pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802": Phase="Pending", Reason="", readiness=false. Elapsed: 27.77109ms Apr 24 23:40:48.878: INFO: Pod "pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03154768s Apr 24 23:40:50.882: INFO: Pod "pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035860689s STEP: Saw pod success Apr 24 23:40:50.882: INFO: Pod "pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802" satisfied condition "Succeeded or Failed" Apr 24 23:40:50.886: INFO: Trying to get logs from node latest-worker pod pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802 container test-container: STEP: delete the pod Apr 24 23:40:50.940: INFO: Waiting for pod pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802 to disappear Apr 24 23:40:50.946: INFO: Pod pod-29cf4eaa-0481-4bcf-88d8-acdab1e54802 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:40:50.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3658" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":22,"skipped":420,"failed":0} SSSSSS ------------------------------ [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:40:50.950: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating all guestbook components Apr 24 23:40:51.009: INFO: apiVersion: v1 kind: Service metadata: name: agnhost-slave labels: app: agnhost role: slave tier: backend spec: ports: - port: 6379 selector: app: agnhost role: slave tier: backend Apr 24 23:40:51.009: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-2098' Apr 24 23:40:53.508: INFO: stderr: "" Apr 24 23:40:53.508: INFO: stdout: "service/agnhost-slave created\n" Apr 24 23:40:53.508: INFO: apiVersion: v1 kind: Service metadata: name: agnhost-master labels: app: agnhost role: master tier: backend spec: ports: - port: 6379 targetPort: 6379 selector: app: agnhost role: master tier: backend Apr 24 23:40:53.508: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-2098' Apr 24 23:40:53.774: INFO: stderr: "" Apr 24 23:40:53.774: INFO: stdout: "service/agnhost-master created\n" Apr 24 23:40:53.774: INFO: apiVersion: v1 kind: Service metadata: name: frontend labels: app: guestbook tier: frontend spec: # if your cluster supports it, uncomment the following to automatically create # an external load-balanced IP for the frontend service. # type: LoadBalancer ports: - port: 80 selector: app: guestbook tier: frontend Apr 24 23:40:53.775: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-2098' Apr 24 23:40:54.081: INFO: stderr: "" Apr 24 23:40:54.081: INFO: stdout: "service/frontend created\n" Apr 24 23:40:54.081: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: frontend spec: replicas: 3 selector: matchLabels: app: guestbook tier: frontend template: metadata: labels: app: guestbook tier: frontend spec: containers: - name: guestbook-frontend image: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 args: [ "guestbook", "--backend-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 80 Apr 24 23:40:54.082: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-2098' Apr 24 23:40:54.374: INFO: stderr: "" Apr 24 23:40:54.374: INFO: stdout: "deployment.apps/frontend created\n" Apr 24 23:40:54.374: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: agnhost-master spec: replicas: 1 selector: matchLabels: app: agnhost role: master tier: backend template: metadata: labels: app: agnhost role: master tier: backend spec: containers: - name: master image: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 args: [ "guestbook", "--http-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Apr 24 23:40:54.374: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-2098' Apr 24 23:40:55.015: INFO: stderr: "" Apr 24 23:40:55.015: INFO: stdout: "deployment.apps/agnhost-master created\n" Apr 24 23:40:55.015: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: agnhost-slave spec: replicas: 2 selector: matchLabels: app: agnhost role: slave tier: backend template: metadata: labels: app: agnhost role: slave tier: backend spec: containers: - name: slave image: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 args: [ "guestbook", "--slaveof", "agnhost-master", "--http-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Apr 24 23:40:55.015: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-2098' Apr 24 23:40:55.635: INFO: stderr: "" Apr 24 23:40:55.635: INFO: stdout: "deployment.apps/agnhost-slave created\n" STEP: validating guestbook app Apr 24 23:40:55.635: INFO: Waiting for all frontend pods to be Running. Apr 24 23:41:05.685: INFO: Waiting for frontend to serve content. Apr 24 23:41:05.696: INFO: Trying to add a new entry to the guestbook. Apr 24 23:41:05.708: INFO: Verifying that added entry can be retrieved. STEP: using delete to clean up resources Apr 24 23:41:05.717: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-2098' Apr 24 23:41:05.870: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 24 23:41:05.871: INFO: stdout: "service \"agnhost-slave\" force deleted\n" STEP: using delete to clean up resources Apr 24 23:41:05.871: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-2098' Apr 24 23:41:06.015: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 24 23:41:06.015: INFO: stdout: "service \"agnhost-master\" force deleted\n" STEP: using delete to clean up resources Apr 24 23:41:06.015: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-2098' Apr 24 23:41:06.172: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 24 23:41:06.172: INFO: stdout: "service \"frontend\" force deleted\n" STEP: using delete to clean up resources Apr 24 23:41:06.172: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-2098' Apr 24 23:41:06.269: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 24 23:41:06.269: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" STEP: using delete to clean up resources Apr 24 23:41:06.269: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-2098' Apr 24 23:41:06.389: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 24 23:41:06.389: INFO: stdout: "deployment.apps \"agnhost-master\" force deleted\n" STEP: using delete to clean up resources Apr 24 23:41:06.390: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-2098' Apr 24 23:41:06.535: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 24 23:41:06.535: INFO: stdout: "deployment.apps \"agnhost-slave\" force deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:41:06.535: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-2098" for this suite. • [SLOW TEST:15.599 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Guestbook application /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:310 should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance]","total":275,"completed":23,"skipped":426,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:41:06.550: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:41:13.666: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-8800" for this suite. • [SLOW TEST:7.126 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","total":275,"completed":24,"skipped":449,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:41:13.676: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Cleaning up the secret STEP: Cleaning up the configmap STEP: Cleaning up the pod [AfterEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:41:17.805: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-6178" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":275,"completed":25,"skipped":472,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:41:17.837: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name cm-test-opt-del-3cccee51-e60d-4044-a3ec-9d2fc76d602a STEP: Creating configMap with name cm-test-opt-upd-7b4a1c29-becf-4d22-97ce-e31d18289aac STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-3cccee51-e60d-4044-a3ec-9d2fc76d602a STEP: Updating configmap cm-test-opt-upd-7b4a1c29-becf-4d22-97ce-e31d18289aac STEP: Creating configMap with name cm-test-opt-create-832948fe-9fbf-40a2-b690-74a4c5f37d1b STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:42:54.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1007" for this suite. • [SLOW TEST:96.936 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:36 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":275,"completed":26,"skipped":490,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:42:54.774: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:74 [It] deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 24 23:42:54.841: INFO: Pod name cleanup-pod: Found 0 pods out of 1 Apr 24 23:42:59.846: INFO: Pod name cleanup-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Apr 24 23:42:59.846: INFO: Creating deployment test-cleanup-deployment STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:68 Apr 24 23:43:03.889: INFO: Deployment "test-cleanup-deployment": &Deployment{ObjectMeta:{test-cleanup-deployment deployment-3354 /apis/apps/v1/namespaces/deployment-3354/deployments/test-cleanup-deployment 40f7be3b-c6da-461b-9990-d4a0282ecc4c 10778434 1 2020-04-24 23:42:59 +0000 UTC map[name:cleanup-pod] map[deployment.kubernetes.io/revision:1] [] [] []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc003d423e8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-04-24 23:42:59 +0000 UTC,LastTransitionTime:2020-04-24 23:42:59 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-cleanup-deployment-577c77b589" has successfully progressed.,LastUpdateTime:2020-04-24 23:43:03 +0000 UTC,LastTransitionTime:2020-04-24 23:42:59 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Apr 24 23:43:03.892: INFO: New ReplicaSet "test-cleanup-deployment-577c77b589" of Deployment "test-cleanup-deployment": &ReplicaSet{ObjectMeta:{test-cleanup-deployment-577c77b589 deployment-3354 /apis/apps/v1/namespaces/deployment-3354/replicasets/test-cleanup-deployment-577c77b589 d8af71f0-5431-4cf0-9781-998d3a0c3a94 10778422 1 2020-04-24 23:42:59 +0000 UTC map[name:cleanup-pod pod-template-hash:577c77b589] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-cleanup-deployment 40f7be3b-c6da-461b-9990-d4a0282ecc4c 0xc003d42837 0xc003d42838}] [] []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 577c77b589,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod pod-template-hash:577c77b589] map[] [] [] []} {[] [] [{agnhost us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc003d428a8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Apr 24 23:43:03.894: INFO: Pod "test-cleanup-deployment-577c77b589-8t26f" is available: &Pod{ObjectMeta:{test-cleanup-deployment-577c77b589-8t26f test-cleanup-deployment-577c77b589- deployment-3354 /api/v1/namespaces/deployment-3354/pods/test-cleanup-deployment-577c77b589-8t26f 13c44267-6cfe-4a6a-b7b8-c0e880eb1e2b 10778421 0 2020-04-24 23:42:59 +0000 UTC map[name:cleanup-pod pod-template-hash:577c77b589] map[] [{apps/v1 ReplicaSet test-cleanup-deployment-577c77b589 d8af71f0-5431-4cf0-9781-998d3a0c3a94 0xc003d42ca7 0xc003d42ca8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-9d8kq,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-9d8kq,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-9d8kq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-24 23:42:59 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-24 23:43:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-24 23:43:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-24 23:42:59 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:10.244.1.126,StartTime:2020-04-24 23:42:59 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-24 23:43:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,ImageID:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost@sha256:1d7f0d77a6f07fd507f147a38d06a7c8269ebabd4f923bfe46d4fb8b396a520c,ContainerID:containerd://797c3e0bb6c7a4752315581248a5cf4c15f42cd16220a031afd6fb560a098557,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.126,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:43:03.894: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-3354" for this suite. • [SLOW TEST:9.126 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":275,"completed":27,"skipped":509,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:43:03.902: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating 50 configmaps STEP: Creating RC which spawns configmap-volume pods Apr 24 23:43:04.731: INFO: Pod name wrapped-volume-race-79115b48-8ca4-489b-9325-80fad70a1763: Found 0 pods out of 5 Apr 24 23:43:10.427: INFO: Pod name wrapped-volume-race-79115b48-8ca4-489b-9325-80fad70a1763: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-79115b48-8ca4-489b-9325-80fad70a1763 in namespace emptydir-wrapper-5826, will wait for the garbage collector to delete the pods Apr 24 23:43:23.288: INFO: Deleting ReplicationController wrapped-volume-race-79115b48-8ca4-489b-9325-80fad70a1763 took: 6.956366ms Apr 24 23:43:23.388: INFO: Terminating ReplicationController wrapped-volume-race-79115b48-8ca4-489b-9325-80fad70a1763 pods took: 100.268299ms STEP: Creating RC which spawns configmap-volume pods Apr 24 23:43:33.844: INFO: Pod name wrapped-volume-race-3d0c479e-3e0b-481d-b592-5fa93fad73ac: Found 0 pods out of 5 Apr 24 23:43:38.873: INFO: Pod name wrapped-volume-race-3d0c479e-3e0b-481d-b592-5fa93fad73ac: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-3d0c479e-3e0b-481d-b592-5fa93fad73ac in namespace emptydir-wrapper-5826, will wait for the garbage collector to delete the pods Apr 24 23:43:52.963: INFO: Deleting ReplicationController wrapped-volume-race-3d0c479e-3e0b-481d-b592-5fa93fad73ac took: 7.285149ms Apr 24 23:43:53.363: INFO: Terminating ReplicationController wrapped-volume-race-3d0c479e-3e0b-481d-b592-5fa93fad73ac pods took: 400.539723ms STEP: Creating RC which spawns configmap-volume pods Apr 24 23:44:03.187: INFO: Pod name wrapped-volume-race-35dee5f1-da5b-426b-8021-5b2358995e76: Found 0 pods out of 5 Apr 24 23:44:08.195: INFO: Pod name wrapped-volume-race-35dee5f1-da5b-426b-8021-5b2358995e76: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-35dee5f1-da5b-426b-8021-5b2358995e76 in namespace emptydir-wrapper-5826, will wait for the garbage collector to delete the pods Apr 24 23:44:22.288: INFO: Deleting ReplicationController wrapped-volume-race-35dee5f1-da5b-426b-8021-5b2358995e76 took: 7.165668ms Apr 24 23:44:22.588: INFO: Terminating ReplicationController wrapped-volume-race-35dee5f1-da5b-426b-8021-5b2358995e76 pods took: 300.262749ms STEP: Cleaning up the configMaps [AfterEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:44:33.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-5826" for this suite. • [SLOW TEST:89.640 seconds] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","total":275,"completed":28,"skipped":556,"failed":0} SS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:44:33.542: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a pod. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Pod that fits quota STEP: Ensuring ResourceQuota status captures the pod usage STEP: Not allowing a pod to be created that exceeds remaining quota STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) STEP: Ensuring a pod cannot update its resource requirements STEP: Ensuring attempts to update pod resource requirements did not change quota usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:44:46.669: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-8584" for this suite. • [SLOW TEST:13.136 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a pod. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":275,"completed":29,"skipped":558,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:44:46.679: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of same group but different versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation Apr 24 23:44:46.750: INFO: >>> kubeConfig: /root/.kube/config STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation Apr 24 23:44:58.297: INFO: >>> kubeConfig: /root/.kube/config Apr 24 23:45:01.236: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:10.836: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-620" for this suite. • [SLOW TEST:24.162 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of same group but different versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":275,"completed":30,"skipped":591,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:10.841: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward api env vars Apr 24 23:45:10.906: INFO: Waiting up to 5m0s for pod "downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0" in namespace "downward-api-684" to be "Succeeded or Failed" Apr 24 23:45:10.910: INFO: Pod "downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0": Phase="Pending", Reason="", readiness=false. Elapsed: 3.812318ms Apr 24 23:45:12.914: INFO: Pod "downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008000286s Apr 24 23:45:14.917: INFO: Pod "downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011387727s STEP: Saw pod success Apr 24 23:45:14.917: INFO: Pod "downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0" satisfied condition "Succeeded or Failed" Apr 24 23:45:14.919: INFO: Trying to get logs from node latest-worker2 pod downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0 container dapi-container: STEP: delete the pod Apr 24 23:45:14.986: INFO: Waiting for pod downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0 to disappear Apr 24 23:45:14.993: INFO: Pod downward-api-7d73b88b-09ea-47e4-83c2-f36de9f3b7b0 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:14.993: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-684" for this suite. •{"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","total":275,"completed":31,"skipped":611,"failed":0} SSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:15.000: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test substitution in container's args Apr 24 23:45:15.068: INFO: Waiting up to 5m0s for pod "var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29" in namespace "var-expansion-4832" to be "Succeeded or Failed" Apr 24 23:45:15.071: INFO: Pod "var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29": Phase="Pending", Reason="", readiness=false. Elapsed: 3.441808ms Apr 24 23:45:17.075: INFO: Pod "var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006904668s Apr 24 23:45:19.079: INFO: Pod "var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010857241s STEP: Saw pod success Apr 24 23:45:19.079: INFO: Pod "var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29" satisfied condition "Succeeded or Failed" Apr 24 23:45:19.081: INFO: Trying to get logs from node latest-worker2 pod var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29 container dapi-container: STEP: delete the pod Apr 24 23:45:19.097: INFO: Waiting for pod var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29 to disappear Apr 24 23:45:19.101: INFO: Pod var-expansion-f33ef69b-91e4-4c4d-93b0-07998daadc29 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:19.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-4832" for this suite. •{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":275,"completed":32,"skipped":619,"failed":0} SSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:19.108: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Performing setup for networking test in namespace pod-network-test-8596 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 24 23:45:19.162: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Apr 24 23:45:19.217: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 24 23:45:21.289: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 24 23:45:23.221: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 24 23:45:25.220: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 24 23:45:27.222: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 24 23:45:29.221: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 24 23:45:31.221: INFO: The status of Pod netserver-0 is Running (Ready = true) Apr 24 23:45:31.227: INFO: The status of Pod netserver-1 is Running (Ready = false) Apr 24 23:45:33.231: INFO: The status of Pod netserver-1 is Running (Ready = false) Apr 24 23:45:35.231: INFO: The status of Pod netserver-1 is Running (Ready = false) Apr 24 23:45:37.231: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Apr 24 23:45:41.289: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.2.79:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-8596 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 24 23:45:41.289: INFO: >>> kubeConfig: /root/.kube/config I0424 23:45:41.323190 8 log.go:172] (0xc0028de4d0) (0xc0016b4dc0) Create stream I0424 23:45:41.323223 8 log.go:172] (0xc0028de4d0) (0xc0016b4dc0) Stream added, broadcasting: 1 I0424 23:45:41.324946 8 log.go:172] (0xc0028de4d0) Reply frame received for 1 I0424 23:45:41.324973 8 log.go:172] (0xc0028de4d0) (0xc001de2140) Create stream I0424 23:45:41.324980 8 log.go:172] (0xc0028de4d0) (0xc001de2140) Stream added, broadcasting: 3 I0424 23:45:41.326207 8 log.go:172] (0xc0028de4d0) Reply frame received for 3 I0424 23:45:41.326275 8 log.go:172] (0xc0028de4d0) (0xc001acf680) Create stream I0424 23:45:41.326295 8 log.go:172] (0xc0028de4d0) (0xc001acf680) Stream added, broadcasting: 5 I0424 23:45:41.327354 8 log.go:172] (0xc0028de4d0) Reply frame received for 5 I0424 23:45:41.412142 8 log.go:172] (0xc0028de4d0) Data frame received for 5 I0424 23:45:41.412191 8 log.go:172] (0xc001acf680) (5) Data frame handling I0424 23:45:41.412222 8 log.go:172] (0xc0028de4d0) Data frame received for 3 I0424 23:45:41.412235 8 log.go:172] (0xc001de2140) (3) Data frame handling I0424 23:45:41.412252 8 log.go:172] (0xc001de2140) (3) Data frame sent I0424 23:45:41.412263 8 log.go:172] (0xc0028de4d0) Data frame received for 3 I0424 23:45:41.412273 8 log.go:172] (0xc001de2140) (3) Data frame handling I0424 23:45:41.413742 8 log.go:172] (0xc0028de4d0) Data frame received for 1 I0424 23:45:41.413764 8 log.go:172] (0xc0016b4dc0) (1) Data frame handling I0424 23:45:41.413774 8 log.go:172] (0xc0016b4dc0) (1) Data frame sent I0424 23:45:41.413790 8 log.go:172] (0xc0028de4d0) (0xc0016b4dc0) Stream removed, broadcasting: 1 I0424 23:45:41.413803 8 log.go:172] (0xc0028de4d0) Go away received I0424 23:45:41.413914 8 log.go:172] (0xc0028de4d0) (0xc0016b4dc0) Stream removed, broadcasting: 1 I0424 23:45:41.413942 8 log.go:172] (0xc0028de4d0) (0xc001de2140) Stream removed, broadcasting: 3 I0424 23:45:41.413961 8 log.go:172] (0xc0028de4d0) (0xc001acf680) Stream removed, broadcasting: 5 Apr 24 23:45:41.413: INFO: Found all expected endpoints: [netserver-0] Apr 24 23:45:41.417: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.1.129:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-8596 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 24 23:45:41.417: INFO: >>> kubeConfig: /root/.kube/config I0424 23:45:41.450807 8 log.go:172] (0xc002dc2370) (0xc001de26e0) Create stream I0424 23:45:41.450839 8 log.go:172] (0xc002dc2370) (0xc001de26e0) Stream added, broadcasting: 1 I0424 23:45:41.452705 8 log.go:172] (0xc002dc2370) Reply frame received for 1 I0424 23:45:41.452760 8 log.go:172] (0xc002dc2370) (0xc001de2780) Create stream I0424 23:45:41.452775 8 log.go:172] (0xc002dc2370) (0xc001de2780) Stream added, broadcasting: 3 I0424 23:45:41.454096 8 log.go:172] (0xc002dc2370) Reply frame received for 3 I0424 23:45:41.454147 8 log.go:172] (0xc002dc2370) (0xc0019f6460) Create stream I0424 23:45:41.454165 8 log.go:172] (0xc002dc2370) (0xc0019f6460) Stream added, broadcasting: 5 I0424 23:45:41.455450 8 log.go:172] (0xc002dc2370) Reply frame received for 5 I0424 23:45:41.536583 8 log.go:172] (0xc002dc2370) Data frame received for 3 I0424 23:45:41.536641 8 log.go:172] (0xc001de2780) (3) Data frame handling I0424 23:45:41.536690 8 log.go:172] (0xc001de2780) (3) Data frame sent I0424 23:45:41.536713 8 log.go:172] (0xc002dc2370) Data frame received for 3 I0424 23:45:41.536734 8 log.go:172] (0xc001de2780) (3) Data frame handling I0424 23:45:41.536924 8 log.go:172] (0xc002dc2370) Data frame received for 5 I0424 23:45:41.536951 8 log.go:172] (0xc0019f6460) (5) Data frame handling I0424 23:45:41.538780 8 log.go:172] (0xc002dc2370) Data frame received for 1 I0424 23:45:41.538810 8 log.go:172] (0xc001de26e0) (1) Data frame handling I0424 23:45:41.538836 8 log.go:172] (0xc001de26e0) (1) Data frame sent I0424 23:45:41.538873 8 log.go:172] (0xc002dc2370) (0xc001de26e0) Stream removed, broadcasting: 1 I0424 23:45:41.538924 8 log.go:172] (0xc002dc2370) Go away received I0424 23:45:41.538997 8 log.go:172] (0xc002dc2370) (0xc001de26e0) Stream removed, broadcasting: 1 I0424 23:45:41.539048 8 log.go:172] (0xc002dc2370) (0xc001de2780) Stream removed, broadcasting: 3 I0424 23:45:41.539093 8 log.go:172] (0xc002dc2370) (0xc0019f6460) Stream removed, broadcasting: 5 Apr 24 23:45:41.539: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:41.539: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-8596" for this suite. • [SLOW TEST:22.441 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":33,"skipped":622,"failed":0} SSSSS ------------------------------ [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:41.549: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating secret secrets-939/secret-test-4e4d1d00-902b-458a-acf3-ef4389cf9321 STEP: Creating a pod to test consume secrets Apr 24 23:45:41.619: INFO: Waiting up to 5m0s for pod "pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370" in namespace "secrets-939" to be "Succeeded or Failed" Apr 24 23:45:41.623: INFO: Pod "pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370": Phase="Pending", Reason="", readiness=false. Elapsed: 3.638828ms Apr 24 23:45:43.626: INFO: Pod "pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007161338s Apr 24 23:45:45.631: INFO: Pod "pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011466707s STEP: Saw pod success Apr 24 23:45:45.631: INFO: Pod "pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370" satisfied condition "Succeeded or Failed" Apr 24 23:45:45.634: INFO: Trying to get logs from node latest-worker pod pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370 container env-test: STEP: delete the pod Apr 24 23:45:45.668: INFO: Waiting for pod pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370 to disappear Apr 24 23:45:45.672: INFO: Pod pod-configmaps-54ffb8d5-f096-4c11-8809-6a474be94370 no longer exists [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:45.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-939" for this suite. •{"msg":"PASSED [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":275,"completed":34,"skipped":627,"failed":0} ------------------------------ [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:45.679: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: validating api versions Apr 24 23:45:45.750: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config api-versions' Apr 24 23:45:45.955: INFO: stderr: "" Apr 24 23:45:45.955: INFO: stdout: "admissionregistration.k8s.io/v1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ndiscovery.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:45.955: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7608" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance]","total":275,"completed":35,"skipped":627,"failed":0} ------------------------------ [sig-network] Services should find a service from listing all namespaces [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:45.963: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should find a service from listing all namespaces [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: fetching services [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:46.018: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-939" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 •{"msg":"PASSED [sig-network] Services should find a service from listing all namespaces [Conformance]","total":275,"completed":36,"skipped":627,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:46.025: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 24 23:45:46.110: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1" in namespace "downward-api-1221" to be "Succeeded or Failed" Apr 24 23:45:46.130: INFO: Pod "downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1": Phase="Pending", Reason="", readiness=false. Elapsed: 19.979285ms Apr 24 23:45:48.148: INFO: Pod "downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.038268946s Apr 24 23:45:50.152: INFO: Pod "downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.042213231s STEP: Saw pod success Apr 24 23:45:50.152: INFO: Pod "downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1" satisfied condition "Succeeded or Failed" Apr 24 23:45:50.155: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1 container client-container: STEP: delete the pod Apr 24 23:45:50.227: INFO: Waiting for pod downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1 to disappear Apr 24 23:45:50.234: INFO: Pod downwardapi-volume-f1ea45bc-5ba0-4049-bcac-e47caaec5ab1 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:50.234: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1221" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":37,"skipped":655,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:50.242: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 24 23:45:51.318: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 24 23:45:53.329: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368751, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368751, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368751, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368751, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 24 23:45:56.356: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny custom resource creation, update and deletion [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 24 23:45:56.360: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the custom resource webhook via the AdmissionRegistration API STEP: Creating a custom resource that should be denied by the webhook STEP: Creating a custom resource whose deletion would be denied by the webhook STEP: Updating the custom resource with disallowed data should be denied STEP: Deleting the custom resource should be denied STEP: Remove the offending key and value from the custom resource data STEP: Deleting the updated custom resource should be successful [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:45:57.507: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-4773" for this suite. STEP: Destroying namespace "webhook-4773-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:7.335 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny custom resource creation, update and deletion [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":275,"completed":38,"skipped":670,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:45:57.577: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should be able to change the type from NodePort to ExternalName [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a service nodeport-service with the type=NodePort in namespace services-8220 STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service STEP: creating service externalsvc in namespace services-8220 STEP: creating replication controller externalsvc in namespace services-8220 I0424 23:45:58.362894 8 runners.go:190] Created replication controller with name: externalsvc, namespace: services-8220, replica count: 2 I0424 23:46:01.413658 8 runners.go:190] externalsvc Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0424 23:46:04.413900 8 runners.go:190] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady STEP: changing the NodePort service to type=ExternalName Apr 24 23:46:04.490: INFO: Creating new exec pod Apr 24 23:46:08.520: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-8220 execpod822fk -- /bin/sh -x -c nslookup nodeport-service' Apr 24 23:46:08.760: INFO: stderr: "I0424 23:46:08.644785 336 log.go:172] (0xc0008e80b0) (0xc000508aa0) Create stream\nI0424 23:46:08.644858 336 log.go:172] (0xc0008e80b0) (0xc000508aa0) Stream added, broadcasting: 1\nI0424 23:46:08.648422 336 log.go:172] (0xc0008e80b0) Reply frame received for 1\nI0424 23:46:08.648477 336 log.go:172] (0xc0008e80b0) (0xc000a6a000) Create stream\nI0424 23:46:08.648502 336 log.go:172] (0xc0008e80b0) (0xc000a6a000) Stream added, broadcasting: 3\nI0424 23:46:08.650257 336 log.go:172] (0xc0008e80b0) Reply frame received for 3\nI0424 23:46:08.650323 336 log.go:172] (0xc0008e80b0) (0xc00097e000) Create stream\nI0424 23:46:08.650343 336 log.go:172] (0xc0008e80b0) (0xc00097e000) Stream added, broadcasting: 5\nI0424 23:46:08.651507 336 log.go:172] (0xc0008e80b0) Reply frame received for 5\nI0424 23:46:08.745638 336 log.go:172] (0xc0008e80b0) Data frame received for 5\nI0424 23:46:08.745672 336 log.go:172] (0xc00097e000) (5) Data frame handling\nI0424 23:46:08.745690 336 log.go:172] (0xc00097e000) (5) Data frame sent\n+ nslookup nodeport-service\nI0424 23:46:08.751689 336 log.go:172] (0xc0008e80b0) Data frame received for 3\nI0424 23:46:08.751712 336 log.go:172] (0xc000a6a000) (3) Data frame handling\nI0424 23:46:08.751729 336 log.go:172] (0xc000a6a000) (3) Data frame sent\nI0424 23:46:08.752464 336 log.go:172] (0xc0008e80b0) Data frame received for 3\nI0424 23:46:08.752476 336 log.go:172] (0xc000a6a000) (3) Data frame handling\nI0424 23:46:08.752487 336 log.go:172] (0xc000a6a000) (3) Data frame sent\nI0424 23:46:08.752928 336 log.go:172] (0xc0008e80b0) Data frame received for 5\nI0424 23:46:08.752948 336 log.go:172] (0xc00097e000) (5) Data frame handling\nI0424 23:46:08.752985 336 log.go:172] (0xc0008e80b0) Data frame received for 3\nI0424 23:46:08.752995 336 log.go:172] (0xc000a6a000) (3) Data frame handling\nI0424 23:46:08.754976 336 log.go:172] (0xc0008e80b0) Data frame received for 1\nI0424 23:46:08.754998 336 log.go:172] (0xc000508aa0) (1) Data frame handling\nI0424 23:46:08.755016 336 log.go:172] (0xc000508aa0) (1) Data frame sent\nI0424 23:46:08.755029 336 log.go:172] (0xc0008e80b0) (0xc000508aa0) Stream removed, broadcasting: 1\nI0424 23:46:08.755051 336 log.go:172] (0xc0008e80b0) Go away received\nI0424 23:46:08.755297 336 log.go:172] (0xc0008e80b0) (0xc000508aa0) Stream removed, broadcasting: 1\nI0424 23:46:08.755325 336 log.go:172] (0xc0008e80b0) (0xc000a6a000) Stream removed, broadcasting: 3\nI0424 23:46:08.755333 336 log.go:172] (0xc0008e80b0) (0xc00097e000) Stream removed, broadcasting: 5\n" Apr 24 23:46:08.761: INFO: stdout: "Server:\t\t10.96.0.10\nAddress:\t10.96.0.10#53\n\nnodeport-service.services-8220.svc.cluster.local\tcanonical name = externalsvc.services-8220.svc.cluster.local.\nName:\texternalsvc.services-8220.svc.cluster.local\nAddress: 10.96.13.242\n\n" STEP: deleting ReplicationController externalsvc in namespace services-8220, will wait for the garbage collector to delete the pods Apr 24 23:46:08.821: INFO: Deleting ReplicationController externalsvc took: 6.517192ms Apr 24 23:46:09.121: INFO: Terminating ReplicationController externalsvc pods took: 300.248354ms Apr 24 23:46:22.846: INFO: Cleaning up the NodePort to ExternalName test service [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:46:22.872: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8220" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 • [SLOW TEST:25.303 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from NodePort to ExternalName [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","total":275,"completed":39,"skipped":693,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:46:22.880: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 24 23:46:23.684: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 24 23:46:25.694: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368783, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368783, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368783, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723368783, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 24 23:46:28.741: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate pod and apply defaults after mutation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering the mutating pod webhook via the AdmissionRegistration API STEP: create a pod that should be updated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:46:29.027: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2710" for this suite. STEP: Destroying namespace "webhook-2710-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:6.337 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate pod and apply defaults after mutation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","total":275,"completed":40,"skipped":705,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:46:29.217: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-upd-9d0d8c93-9b14-4b62-b6d1-0a5ff891759b STEP: Creating the pod STEP: Waiting for pod with text data STEP: Waiting for pod with binary data [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:46:35.476: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8516" for this suite. • [SLOW TEST:6.267 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":275,"completed":41,"skipped":729,"failed":0} SSSSSSSSSS ------------------------------ [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:46:35.484: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name cm-test-opt-del-096555b1-b27a-4801-8d76-2e4be64f265b STEP: Creating configMap with name cm-test-opt-upd-b40b653a-7c84-4e48-90a0-7fc25ac37682 STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-096555b1-b27a-4801-8d76-2e4be64f265b STEP: Updating configmap cm-test-opt-upd-b40b653a-7c84-4e48-90a0-7fc25ac37682 STEP: Creating configMap with name cm-test-opt-create-cba7a6a0-10db-4c5e-a707-5a45a89d8e9f STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:46:45.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-6009" for this suite. • [SLOW TEST:10.218 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":275,"completed":42,"skipped":739,"failed":0} S ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:46:45.702: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 24 23:46:45.757: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8" in namespace "projected-345" to be "Succeeded or Failed" Apr 24 23:46:45.778: INFO: Pod "downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8": Phase="Pending", Reason="", readiness=false. Elapsed: 21.1973ms Apr 24 23:46:47.783: INFO: Pod "downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025551917s Apr 24 23:46:49.786: INFO: Pod "downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028673438s STEP: Saw pod success Apr 24 23:46:49.786: INFO: Pod "downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8" satisfied condition "Succeeded or Failed" Apr 24 23:46:49.788: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8 container client-container: STEP: delete the pod Apr 24 23:46:49.806: INFO: Waiting for pod downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8 to disappear Apr 24 23:46:49.810: INFO: Pod downwardapi-volume-8690fa77-4408-49de-b82c-a616d6c3c2a8 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:46:49.810: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-345" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","total":275,"completed":43,"skipped":740,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] PreStop should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:46:49.818: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename prestop STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:171 [It] should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating server pod server in namespace prestop-5159 STEP: Waiting for pods to come up. STEP: Creating tester pod tester in namespace prestop-5159 STEP: Deleting pre-stop pod Apr 24 23:47:02.931: INFO: Saw: { "Hostname": "server", "Sent": null, "Received": { "prestop": 1 }, "Errors": null, "Log": [ "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." ], "StillContactingPeers": true } STEP: Deleting the server pod [AfterEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:47:02.936: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "prestop-5159" for this suite. • [SLOW TEST:13.156 seconds] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] [sig-node] PreStop should call prestop when killing a pod [Conformance]","total":275,"completed":44,"skipped":806,"failed":0} SSSSS ------------------------------ [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:47:02.975: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:03.027: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-1516" for this suite. • [SLOW TEST:60.060 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","total":275,"completed":45,"skipped":811,"failed":0} SSSSSS ------------------------------ [sig-apps] ReplicationController should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:03.035: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Given a Pod with a 'name' label pod-adoption is created STEP: When a replication controller with a matching selector is created STEP: Then the orphan pod is adopted [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:08.168: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-692" for this suite. • [SLOW TEST:5.141 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","total":275,"completed":46,"skipped":817,"failed":0} SSS ------------------------------ [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:08.176: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 24 23:48:08.234: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4" in namespace "projected-4459" to be "Succeeded or Failed" Apr 24 23:48:08.288: INFO: Pod "downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4": Phase="Pending", Reason="", readiness=false. Elapsed: 54.602647ms Apr 24 23:48:10.292: INFO: Pod "downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.05830324s Apr 24 23:48:12.297: INFO: Pod "downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.063302591s STEP: Saw pod success Apr 24 23:48:12.297: INFO: Pod "downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4" satisfied condition "Succeeded or Failed" Apr 24 23:48:12.300: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4 container client-container: STEP: delete the pod Apr 24 23:48:12.389: INFO: Waiting for pod downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4 to disappear Apr 24 23:48:12.393: INFO: Pod downwardapi-volume-d58082f4-3c52-4d1f-b6bc-01e7aa86a0c4 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:12.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4459" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","total":275,"completed":47,"skipped":820,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:12.399: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 24 23:48:12.462: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310" in namespace "projected-307" to be "Succeeded or Failed" Apr 24 23:48:12.465: INFO: Pod "downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310": Phase="Pending", Reason="", readiness=false. Elapsed: 3.600135ms Apr 24 23:48:14.478: INFO: Pod "downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016751545s Apr 24 23:48:16.482: INFO: Pod "downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02036218s STEP: Saw pod success Apr 24 23:48:16.482: INFO: Pod "downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310" satisfied condition "Succeeded or Failed" Apr 24 23:48:16.485: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310 container client-container: STEP: delete the pod Apr 24 23:48:16.526: INFO: Waiting for pod downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310 to disappear Apr 24 23:48:16.531: INFO: Pod downwardapi-volume-2f8577b4-1eaa-4bc0-b439-4e09e170e310 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:16.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-307" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":275,"completed":48,"skipped":854,"failed":0} SSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:16.556: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD preserving unknown fields at the schema root [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 24 23:48:16.610: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Apr 24 23:48:19.556: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-3744 create -f -' Apr 24 23:48:22.190: INFO: stderr: "" Apr 24 23:48:22.190: INFO: stdout: "e2e-test-crd-publish-openapi-5992-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" Apr 24 23:48:22.190: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-3744 delete e2e-test-crd-publish-openapi-5992-crds test-cr' Apr 24 23:48:22.321: INFO: stderr: "" Apr 24 23:48:22.321: INFO: stdout: "e2e-test-crd-publish-openapi-5992-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" Apr 24 23:48:22.321: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-3744 apply -f -' Apr 24 23:48:22.594: INFO: stderr: "" Apr 24 23:48:22.594: INFO: stdout: "e2e-test-crd-publish-openapi-5992-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" Apr 24 23:48:22.594: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-3744 delete e2e-test-crd-publish-openapi-5992-crds test-cr' Apr 24 23:48:22.716: INFO: stderr: "" Apr 24 23:48:22.716: INFO: stdout: "e2e-test-crd-publish-openapi-5992-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR Apr 24 23:48:22.716: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-5992-crds' Apr 24 23:48:22.962: INFO: stderr: "" Apr 24 23:48:22.962: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-5992-crd\nVERSION: crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n \n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:25.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-3744" for this suite. • [SLOW TEST:9.333 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD preserving unknown fields at the schema root [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","total":275,"completed":49,"skipped":867,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:25.889: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a watch on configmaps STEP: creating a new configmap STEP: modifying the configmap once STEP: closing the watch once it receives two notifications Apr 24 23:48:25.975: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4381 /api/v1/namespaces/watch-4381/configmaps/e2e-watch-test-watch-closed 13b68335-3ef3-40db-8f7c-3b565adfd6d7 10780991 0 2020-04-24 23:48:25 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Apr 24 23:48:25.976: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4381 /api/v1/namespaces/watch-4381/configmaps/e2e-watch-test-watch-closed 13b68335-3ef3-40db-8f7c-3b565adfd6d7 10780992 0 2020-04-24 23:48:25 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying the configmap a second time, while the watch is closed STEP: creating a new watch on configmaps from the last resource version observed by the first watch STEP: deleting the configmap STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed Apr 24 23:48:25.988: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4381 /api/v1/namespaces/watch-4381/configmaps/e2e-watch-test-watch-closed 13b68335-3ef3-40db-8f7c-3b565adfd6d7 10780993 0 2020-04-24 23:48:25 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 24 23:48:25.988: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4381 /api/v1/namespaces/watch-4381/configmaps/e2e-watch-test-watch-closed 13b68335-3ef3-40db-8f7c-3b565adfd6d7 10780994 0 2020-04-24 23:48:25 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:25.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-4381" for this suite. •{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":275,"completed":50,"skipped":885,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:25.998: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating projection with secret that has name projected-secret-test-map-42e151a7-aaed-42e1-8863-c75a6e4924a9 STEP: Creating a pod to test consume secrets Apr 24 23:48:26.092: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d" in namespace "projected-3488" to be "Succeeded or Failed" Apr 24 23:48:26.100: INFO: Pod "pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d": Phase="Pending", Reason="", readiness=false. Elapsed: 8.210894ms Apr 24 23:48:28.137: INFO: Pod "pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045023153s Apr 24 23:48:30.141: INFO: Pod "pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04951926s STEP: Saw pod success Apr 24 23:48:30.141: INFO: Pod "pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d" satisfied condition "Succeeded or Failed" Apr 24 23:48:30.145: INFO: Trying to get logs from node latest-worker2 pod pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d container projected-secret-volume-test: STEP: delete the pod Apr 24 23:48:30.163: INFO: Waiting for pod pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d to disappear Apr 24 23:48:30.166: INFO: Pod pod-projected-secrets-d5bedcd8-904e-4707-b299-b8332c261e1d no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:30.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3488" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":275,"completed":51,"skipped":929,"failed":0} SSSSSSS ------------------------------ [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:30.173: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-projected-all-test-volume-18f63e85-0278-4d11-93fa-7b8064871528 STEP: Creating secret with name secret-projected-all-test-volume-712f3425-5e15-457d-a385-18e2a3119d09 STEP: Creating a pod to test Check all projections for projected volume plugin Apr 24 23:48:30.318: INFO: Waiting up to 5m0s for pod "projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1" in namespace "projected-470" to be "Succeeded or Failed" Apr 24 23:48:30.322: INFO: Pod "projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1": Phase="Pending", Reason="", readiness=false. Elapsed: 3.748719ms Apr 24 23:48:32.677: INFO: Pod "projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.359007784s Apr 24 23:48:34.682: INFO: Pod "projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.363231097s STEP: Saw pod success Apr 24 23:48:34.682: INFO: Pod "projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1" satisfied condition "Succeeded or Failed" Apr 24 23:48:34.685: INFO: Trying to get logs from node latest-worker2 pod projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1 container projected-all-volume-test: STEP: delete the pod Apr 24 23:48:34.723: INFO: Waiting for pod projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1 to disappear Apr 24 23:48:34.734: INFO: Pod projected-volume-b8ab4cdc-2b4c-4477-a3b5-c8be518ec7d1 no longer exists [AfterEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:34.734: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-470" for this suite. •{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":275,"completed":52,"skipped":936,"failed":0} SSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:34.741: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] listing custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 24 23:48:34.798: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:41.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-9594" for this suite. • [SLOW TEST:6.289 seconds] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Simple CustomResourceDefinition /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:48 listing custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance]","total":275,"completed":53,"skipped":939,"failed":0} [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:41.030: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-volume-map-0412679c-6999-4906-a833-9aff92f0c9bd STEP: Creating a pod to test consume configMaps Apr 24 23:48:41.097: INFO: Waiting up to 5m0s for pod "pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db" in namespace "configmap-5273" to be "Succeeded or Failed" Apr 24 23:48:41.144: INFO: Pod "pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db": Phase="Pending", Reason="", readiness=false. Elapsed: 46.958106ms Apr 24 23:48:43.147: INFO: Pod "pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db": Phase="Pending", Reason="", readiness=false. Elapsed: 2.050708854s Apr 24 23:48:45.152: INFO: Pod "pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.054837346s STEP: Saw pod success Apr 24 23:48:45.152: INFO: Pod "pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db" satisfied condition "Succeeded or Failed" Apr 24 23:48:45.155: INFO: Trying to get logs from node latest-worker pod pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db container configmap-volume-test: STEP: delete the pod Apr 24 23:48:45.181: INFO: Waiting for pod pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db to disappear Apr 24 23:48:45.185: INFO: Pod pod-configmaps-752d551f-f58f-45fd-8f59-166f506d81db no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:45.185: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5273" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":275,"completed":54,"skipped":939,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:45.192: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating replication controller my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955 Apr 24 23:48:45.278: INFO: Pod name my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955: Found 0 pods out of 1 Apr 24 23:48:50.300: INFO: Pod name my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955: Found 1 pods out of 1 Apr 24 23:48:50.300: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955" are running Apr 24 23:48:50.316: INFO: Pod "my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955-mq5cf" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-24 23:48:45 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-24 23:48:48 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-24 23:48:48 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-24 23:48:45 +0000 UTC Reason: Message:}]) Apr 24 23:48:50.316: INFO: Trying to dial the pod Apr 24 23:48:55.328: INFO: Controller my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955: Got expected result from replica 1 [my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955-mq5cf]: "my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955-mq5cf", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:55.328: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-5964" for this suite. • [SLOW TEST:10.142 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance]","total":275,"completed":55,"skipped":958,"failed":0} SSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:55.335: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0777 on node default medium Apr 24 23:48:55.427: INFO: Waiting up to 5m0s for pod "pod-929ef8ff-74fd-4848-9e9d-847988d9c959" in namespace "emptydir-3823" to be "Succeeded or Failed" Apr 24 23:48:55.442: INFO: Pod "pod-929ef8ff-74fd-4848-9e9d-847988d9c959": Phase="Pending", Reason="", readiness=false. Elapsed: 14.440791ms Apr 24 23:48:57.445: INFO: Pod "pod-929ef8ff-74fd-4848-9e9d-847988d9c959": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017842158s Apr 24 23:48:59.449: INFO: Pod "pod-929ef8ff-74fd-4848-9e9d-847988d9c959": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021336911s STEP: Saw pod success Apr 24 23:48:59.449: INFO: Pod "pod-929ef8ff-74fd-4848-9e9d-847988d9c959" satisfied condition "Succeeded or Failed" Apr 24 23:48:59.451: INFO: Trying to get logs from node latest-worker pod pod-929ef8ff-74fd-4848-9e9d-847988d9c959 container test-container: STEP: delete the pod Apr 24 23:48:59.480: INFO: Waiting for pod pod-929ef8ff-74fd-4848-9e9d-847988d9c959 to disappear Apr 24 23:48:59.520: INFO: Pod pod-929ef8ff-74fd-4848-9e9d-847988d9c959 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:48:59.520: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3823" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":56,"skipped":971,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:48:59.529: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: getting the auto-created API token Apr 24 23:49:00.183: INFO: created pod pod-service-account-defaultsa Apr 24 23:49:00.183: INFO: pod pod-service-account-defaultsa service account token volume mount: true Apr 24 23:49:00.191: INFO: created pod pod-service-account-mountsa Apr 24 23:49:00.191: INFO: pod pod-service-account-mountsa service account token volume mount: true Apr 24 23:49:00.241: INFO: created pod pod-service-account-nomountsa Apr 24 23:49:00.241: INFO: pod pod-service-account-nomountsa service account token volume mount: false Apr 24 23:49:00.266: INFO: created pod pod-service-account-defaultsa-mountspec Apr 24 23:49:00.266: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true Apr 24 23:49:00.303: INFO: created pod pod-service-account-mountsa-mountspec Apr 24 23:49:00.303: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true Apr 24 23:49:00.317: INFO: created pod pod-service-account-nomountsa-mountspec Apr 24 23:49:00.317: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true Apr 24 23:49:00.380: INFO: created pod pod-service-account-defaultsa-nomountspec Apr 24 23:49:00.380: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false Apr 24 23:49:00.387: INFO: created pod pod-service-account-mountsa-nomountspec Apr 24 23:49:00.387: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false Apr 24 23:49:00.432: INFO: created pod pod-service-account-nomountsa-nomountspec Apr 24 23:49:00.432: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false [AfterEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:49:00.432: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-247" for this suite. •{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]","total":275,"completed":57,"skipped":986,"failed":0} SSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:49:00.559: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:91 Apr 24 23:49:00.674: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 24 23:49:00.696: INFO: Waiting for terminating namespaces to be deleted... Apr 24 23:49:00.698: INFO: Logging pods the kubelet thinks is on node latest-worker before test Apr 24 23:49:00.704: INFO: pod-service-account-nomountsa from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.704: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.704: INFO: pod-service-account-mountsa from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.704: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.704: INFO: pod-service-account-mountsa-mountspec from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.704: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.704: INFO: kindnet-vnjgh from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.704: INFO: Container kindnet-cni ready: true, restart count 0 Apr 24 23:49:00.704: INFO: pod-service-account-nomountsa-mountspec from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.704: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.704: INFO: kube-proxy-s9v6p from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.704: INFO: Container kube-proxy ready: true, restart count 0 Apr 24 23:49:00.704: INFO: pod-service-account-nomountsa-nomountspec from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.704: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.704: INFO: Logging pods the kubelet thinks is on node latest-worker2 before test Apr 24 23:49:00.709: INFO: pod-service-account-mountsa-nomountspec from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.709: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.709: INFO: pod-service-account-defaultsa from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.709: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.709: INFO: pod-service-account-defaultsa-nomountspec from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.709: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.709: INFO: my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955-mq5cf from replication-controller-5964 started at 2020-04-24 23:48:45 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.709: INFO: Container my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955 ready: true, restart count 0 Apr 24 23:49:00.709: INFO: pod-service-account-defaultsa-mountspec from svcaccounts-247 started at 2020-04-24 23:49:00 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.709: INFO: Container token-test ready: false, restart count 0 Apr 24 23:49:00.709: INFO: kindnet-zq6gp from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.709: INFO: Container kindnet-cni ready: true, restart count 0 Apr 24 23:49:00.709: INFO: kube-proxy-c5xlk from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 24 23:49:00.709: INFO: Container kube-proxy ready: true, restart count 0 [It] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: verifying the node has the label node latest-worker STEP: verifying the node has the label node latest-worker2 Apr 24 23:49:00.799: INFO: Pod kindnet-vnjgh requesting resource cpu=100m on Node latest-worker Apr 24 23:49:00.799: INFO: Pod kindnet-zq6gp requesting resource cpu=100m on Node latest-worker2 Apr 24 23:49:00.799: INFO: Pod kube-proxy-c5xlk requesting resource cpu=0m on Node latest-worker2 Apr 24 23:49:00.799: INFO: Pod kube-proxy-s9v6p requesting resource cpu=0m on Node latest-worker Apr 24 23:49:00.799: INFO: Pod my-hostname-basic-713dce16-897b-4afd-9e63-beda2aa03955-mq5cf requesting resource cpu=0m on Node latest-worker2 Apr 24 23:49:00.799: INFO: Pod pod-service-account-defaultsa requesting resource cpu=0m on Node latest-worker2 Apr 24 23:49:00.799: INFO: Pod pod-service-account-defaultsa-mountspec requesting resource cpu=0m on Node latest-worker2 Apr 24 23:49:00.799: INFO: Pod pod-service-account-defaultsa-nomountspec requesting resource cpu=0m on Node latest-worker2 Apr 24 23:49:00.799: INFO: Pod pod-service-account-mountsa requesting resource cpu=0m on Node latest-worker Apr 24 23:49:00.799: INFO: Pod pod-service-account-mountsa-mountspec requesting resource cpu=0m on Node latest-worker Apr 24 23:49:00.799: INFO: Pod pod-service-account-mountsa-nomountspec requesting resource cpu=0m on Node latest-worker2 Apr 24 23:49:00.799: INFO: Pod pod-service-account-nomountsa requesting resource cpu=0m on Node latest-worker Apr 24 23:49:00.799: INFO: Pod pod-service-account-nomountsa-mountspec requesting resource cpu=0m on Node latest-worker Apr 24 23:49:00.799: INFO: Pod pod-service-account-nomountsa-nomountspec requesting resource cpu=0m on Node latest-worker STEP: Starting Pods to consume most of the cluster CPU. Apr 24 23:49:00.799: INFO: Creating a pod which consumes cpu=11130m on Node latest-worker Apr 24 23:49:00.805: INFO: Creating a pod which consumes cpu=11130m on Node latest-worker2 STEP: Creating another pod that requires unavailable amount of CPU. STEP: Considering event: Type = [Normal], Name = [filler-pod-002b5f98-20fb-49f2-96ee-87e5cc1992c8.1608e65987b3e6ee], Reason = [Scheduled], Message = [Successfully assigned sched-pred-2000/filler-pod-002b5f98-20fb-49f2-96ee-87e5cc1992c8 to latest-worker2] STEP: Considering event: Type = [Normal], Name = [filler-pod-002b5f98-20fb-49f2-96ee-87e5cc1992c8.1608e65b04d51614], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.2" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-002b5f98-20fb-49f2-96ee-87e5cc1992c8.1608e65be76b648f], Reason = [Created], Message = [Created container filler-pod-002b5f98-20fb-49f2-96ee-87e5cc1992c8] STEP: Considering event: Type = [Normal], Name = [filler-pod-002b5f98-20fb-49f2-96ee-87e5cc1992c8.1608e65c10f5fb13], Reason = [Started], Message = [Started container filler-pod-002b5f98-20fb-49f2-96ee-87e5cc1992c8] STEP: Considering event: Type = [Normal], Name = [filler-pod-7401c068-65de-41cc-91e6-40db1c9c7bce.1608e65985e4986e], Reason = [Scheduled], Message = [Successfully assigned sched-pred-2000/filler-pod-7401c068-65de-41cc-91e6-40db1c9c7bce to latest-worker] STEP: Considering event: Type = [Normal], Name = [filler-pod-7401c068-65de-41cc-91e6-40db1c9c7bce.1608e65b585c500e], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.2" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-7401c068-65de-41cc-91e6-40db1c9c7bce.1608e65c37feb21c], Reason = [Created], Message = [Created container filler-pod-7401c068-65de-41cc-91e6-40db1c9c7bce] STEP: Considering event: Type = [Normal], Name = [filler-pod-7401c068-65de-41cc-91e6-40db1c9c7bce.1608e65c4bebe39c], Reason = [Started], Message = [Started container filler-pod-7401c068-65de-41cc-91e6-40db1c9c7bce] STEP: Considering event: Type = [Warning], Name = [additional-pod.1608e65c55ec07f8], Reason = [FailedScheduling], Message = [0/3 nodes are available: 1 node(s) had taints that the pod didn't tolerate, 2 Insufficient cpu.] STEP: removing the label node off the node latest-worker2 STEP: verifying the node doesn't have the label node STEP: removing the label node off the node latest-worker STEP: verifying the node doesn't have the label node [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:49:13.956: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-2000" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:82 • [SLOW TEST:13.404 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance]","total":275,"completed":58,"skipped":990,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:49:13.964: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should be able to change the type from ExternalName to ClusterIP [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a service externalname-service with the type=ExternalName in namespace services-4449 STEP: changing the ExternalName service to type=ClusterIP STEP: creating replication controller externalname-service in namespace services-4449 I0424 23:49:14.073261 8 runners.go:190] Created replication controller with name: externalname-service, namespace: services-4449, replica count: 2 I0424 23:49:17.123766 8 runners.go:190] externalname-service Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0424 23:49:20.123993 8 runners.go:190] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Apr 24 23:49:20.124: INFO: Creating new exec pod Apr 24 23:49:25.136: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-4449 execpodpzjhz -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80' Apr 24 23:49:25.341: INFO: stderr: "I0424 23:49:25.271614 471 log.go:172] (0xc000adcfd0) (0xc000a98640) Create stream\nI0424 23:49:25.271670 471 log.go:172] (0xc000adcfd0) (0xc000a98640) Stream added, broadcasting: 1\nI0424 23:49:25.276643 471 log.go:172] (0xc000adcfd0) Reply frame received for 1\nI0424 23:49:25.276682 471 log.go:172] (0xc000adcfd0) (0xc0007e77c0) Create stream\nI0424 23:49:25.276691 471 log.go:172] (0xc000adcfd0) (0xc0007e77c0) Stream added, broadcasting: 3\nI0424 23:49:25.277889 471 log.go:172] (0xc000adcfd0) Reply frame received for 3\nI0424 23:49:25.277942 471 log.go:172] (0xc000adcfd0) (0xc000540be0) Create stream\nI0424 23:49:25.277959 471 log.go:172] (0xc000adcfd0) (0xc000540be0) Stream added, broadcasting: 5\nI0424 23:49:25.278966 471 log.go:172] (0xc000adcfd0) Reply frame received for 5\nI0424 23:49:25.334307 471 log.go:172] (0xc000adcfd0) Data frame received for 5\nI0424 23:49:25.334355 471 log.go:172] (0xc000540be0) (5) Data frame handling\nI0424 23:49:25.334385 471 log.go:172] (0xc000540be0) (5) Data frame sent\n+ nc -zv -t -w 2 externalname-service 80\nI0424 23:49:25.334668 471 log.go:172] (0xc000adcfd0) Data frame received for 5\nI0424 23:49:25.334772 471 log.go:172] (0xc000540be0) (5) Data frame handling\nI0424 23:49:25.334802 471 log.go:172] (0xc000540be0) (5) Data frame sent\nConnection to externalname-service 80 port [tcp/http] succeeded!\nI0424 23:49:25.335076 471 log.go:172] (0xc000adcfd0) Data frame received for 5\nI0424 23:49:25.335096 471 log.go:172] (0xc000540be0) (5) Data frame handling\nI0424 23:49:25.335128 471 log.go:172] (0xc000adcfd0) Data frame received for 3\nI0424 23:49:25.335167 471 log.go:172] (0xc0007e77c0) (3) Data frame handling\nI0424 23:49:25.337064 471 log.go:172] (0xc000adcfd0) Data frame received for 1\nI0424 23:49:25.337094 471 log.go:172] (0xc000a98640) (1) Data frame handling\nI0424 23:49:25.337249 471 log.go:172] (0xc000a98640) (1) Data frame sent\nI0424 23:49:25.337286 471 log.go:172] (0xc000adcfd0) (0xc000a98640) Stream removed, broadcasting: 1\nI0424 23:49:25.337367 471 log.go:172] (0xc000adcfd0) Go away received\nI0424 23:49:25.337754 471 log.go:172] (0xc000adcfd0) (0xc000a98640) Stream removed, broadcasting: 1\nI0424 23:49:25.337786 471 log.go:172] (0xc000adcfd0) (0xc0007e77c0) Stream removed, broadcasting: 3\nI0424 23:49:25.337797 471 log.go:172] (0xc000adcfd0) (0xc000540be0) Stream removed, broadcasting: 5\n" Apr 24 23:49:25.341: INFO: stdout: "" Apr 24 23:49:25.343: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-4449 execpodpzjhz -- /bin/sh -x -c nc -zv -t -w 2 10.96.72.119 80' Apr 24 23:49:25.543: INFO: stderr: "I0424 23:49:25.468835 491 log.go:172] (0xc0009d0840) (0xc00065e1e0) Create stream\nI0424 23:49:25.468891 491 log.go:172] (0xc0009d0840) (0xc00065e1e0) Stream added, broadcasting: 1\nI0424 23:49:25.471488 491 log.go:172] (0xc0009d0840) Reply frame received for 1\nI0424 23:49:25.471548 491 log.go:172] (0xc0009d0840) (0xc0003cf0e0) Create stream\nI0424 23:49:25.471576 491 log.go:172] (0xc0009d0840) (0xc0003cf0e0) Stream added, broadcasting: 3\nI0424 23:49:25.472555 491 log.go:172] (0xc0009d0840) Reply frame received for 3\nI0424 23:49:25.472604 491 log.go:172] (0xc0009d0840) (0xc00065e280) Create stream\nI0424 23:49:25.472619 491 log.go:172] (0xc0009d0840) (0xc00065e280) Stream added, broadcasting: 5\nI0424 23:49:25.473991 491 log.go:172] (0xc0009d0840) Reply frame received for 5\nI0424 23:49:25.536878 491 log.go:172] (0xc0009d0840) Data frame received for 5\nI0424 23:49:25.536940 491 log.go:172] (0xc00065e280) (5) Data frame handling\nI0424 23:49:25.536965 491 log.go:172] (0xc00065e280) (5) Data frame sent\nI0424 23:49:25.536978 491 log.go:172] (0xc0009d0840) Data frame received for 5\nI0424 23:49:25.536989 491 log.go:172] (0xc00065e280) (5) Data frame handling\n+ nc -zv -t -w 2 10.96.72.119 80\nConnection to 10.96.72.119 80 port [tcp/http] succeeded!\nI0424 23:49:25.537038 491 log.go:172] (0xc0009d0840) Data frame received for 3\nI0424 23:49:25.537070 491 log.go:172] (0xc0003cf0e0) (3) Data frame handling\nI0424 23:49:25.538978 491 log.go:172] (0xc0009d0840) Data frame received for 1\nI0424 23:49:25.539007 491 log.go:172] (0xc00065e1e0) (1) Data frame handling\nI0424 23:49:25.539018 491 log.go:172] (0xc00065e1e0) (1) Data frame sent\nI0424 23:49:25.539033 491 log.go:172] (0xc0009d0840) (0xc00065e1e0) Stream removed, broadcasting: 1\nI0424 23:49:25.539293 491 log.go:172] (0xc0009d0840) (0xc00065e1e0) Stream removed, broadcasting: 1\nI0424 23:49:25.539309 491 log.go:172] (0xc0009d0840) (0xc0003cf0e0) Stream removed, broadcasting: 3\nI0424 23:49:25.539343 491 log.go:172] (0xc0009d0840) Go away received\nI0424 23:49:25.539408 491 log.go:172] (0xc0009d0840) (0xc00065e280) Stream removed, broadcasting: 5\n" Apr 24 23:49:25.543: INFO: stdout: "" Apr 24 23:49:25.543: INFO: Cleaning up the ExternalName to ClusterIP test service [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:49:25.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-4449" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 • [SLOW TEST:11.660 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from ExternalName to ClusterIP [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","total":275,"completed":59,"skipped":1004,"failed":0} SSSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:49:25.624: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-watch STEP: Waiting for a default service account to be provisioned in namespace [It] watch on custom resource definition objects [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 24 23:49:25.664: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating first CR Apr 24 23:49:26.307: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-04-24T23:49:26Z generation:1 name:name1 resourceVersion:10781567 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:eb2af2da-d39b-4b2f-94f5-b78bcb2acbc2] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Creating second CR Apr 24 23:49:36.312: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-04-24T23:49:36Z generation:1 name:name2 resourceVersion:10781622 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:07347858-3962-45e0-9fe4-f605d830019a] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Modifying first CR Apr 24 23:49:46.318: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-04-24T23:49:26Z generation:2 name:name1 resourceVersion:10781652 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:eb2af2da-d39b-4b2f-94f5-b78bcb2acbc2] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Modifying second CR Apr 24 23:49:56.324: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-04-24T23:49:36Z generation:2 name:name2 resourceVersion:10781682 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:07347858-3962-45e0-9fe4-f605d830019a] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Deleting first CR Apr 24 23:50:06.333: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-04-24T23:49:26Z generation:2 name:name1 resourceVersion:10781712 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:eb2af2da-d39b-4b2f-94f5-b78bcb2acbc2] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Deleting second CR Apr 24 23:50:16.339: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-04-24T23:49:36Z generation:2 name:name2 resourceVersion:10781742 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:07347858-3962-45e0-9fe4-f605d830019a] num:map[num1:9223372036854775807 num2:1000000]]} [AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:50:26.852: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-watch-9463" for this suite. • [SLOW TEST:61.238 seconds] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 CustomResourceDefinition Watch /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_watch.go:42 watch on custom resource definition objects [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","total":275,"completed":60,"skipped":1009,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:50:26.862: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-cae84883-f58e-4fa3-a8b3-90cbaa2100fd STEP: Creating a pod to test consume secrets Apr 24 23:50:26.918: INFO: Waiting up to 5m0s for pod "pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9" in namespace "secrets-8795" to be "Succeeded or Failed" Apr 24 23:50:26.923: INFO: Pod "pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.381013ms Apr 24 23:50:29.116: INFO: Pod "pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.197139351s Apr 24 23:50:31.122: INFO: Pod "pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.203320126s STEP: Saw pod success Apr 24 23:50:31.122: INFO: Pod "pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9" satisfied condition "Succeeded or Failed" Apr 24 23:50:31.125: INFO: Trying to get logs from node latest-worker pod pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9 container secret-volume-test: STEP: delete the pod Apr 24 23:50:31.255: INFO: Waiting for pod pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9 to disappear Apr 24 23:50:31.289: INFO: Pod pod-secrets-a81e3eb0-32a7-40e7-85e8-12e00201aee9 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:50:31.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-8795" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":61,"skipped":1044,"failed":0} SSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:50:31.298: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [BeforeEach] when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:82 [It] should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:50:31.530: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-7715" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":275,"completed":62,"skipped":1047,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:50:31.540: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 24 23:50:32.278: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 24 23:50:34.289: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723369032, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723369032, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723369032, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723369032, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 24 23:50:37.303: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should unconditionally reject operations on fail closed webhook [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API STEP: create a namespace for the webhook STEP: create a configmap should be unconditionally rejected by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:50:37.363: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-9739" for this suite. STEP: Destroying namespace "webhook-9739-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:5.912 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should unconditionally reject operations on fail closed webhook [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":275,"completed":63,"skipped":1055,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:50:37.453: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name projected-secret-test-95b09469-09bb-4be8-adaa-9b4492122ff0 STEP: Creating a pod to test consume secrets Apr 24 23:50:37.553: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff" in namespace "projected-302" to be "Succeeded or Failed" Apr 24 23:50:37.580: INFO: Pod "pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff": Phase="Pending", Reason="", readiness=false. Elapsed: 27.286999ms Apr 24 23:50:39.585: INFO: Pod "pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031700182s Apr 24 23:50:41.589: INFO: Pod "pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.03581712s STEP: Saw pod success Apr 24 23:50:41.589: INFO: Pod "pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff" satisfied condition "Succeeded or Failed" Apr 24 23:50:41.592: INFO: Trying to get logs from node latest-worker pod pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff container secret-volume-test: STEP: delete the pod Apr 24 23:50:41.619: INFO: Waiting for pod pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff to disappear Apr 24 23:50:41.648: INFO: Pod pod-projected-secrets-5c6e6467-4365-4c0a-8e37-3b12ba2b87ff no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:50:41.649: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-302" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":275,"completed":64,"skipped":1090,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:50:41.657: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name s-test-opt-del-30d7c6be-4062-4f25-970e-37a9ad66cd02 STEP: Creating secret with name s-test-opt-upd-6e0e0e97-14d8-49b0-821f-07a6a38c9dce STEP: Creating the pod STEP: Deleting secret s-test-opt-del-30d7c6be-4062-4f25-970e-37a9ad66cd02 STEP: Updating secret s-test-opt-upd-6e0e0e97-14d8-49b0-821f-07a6a38c9dce STEP: Creating secret with name s-test-opt-create-d2c79761-b119-40a2-a559-e4abbd4cb91d STEP: waiting to observe update in volume [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:52:16.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-1297" for this suite. • [SLOW TEST:94.629 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:36 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","total":275,"completed":65,"skipped":1132,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:52:16.287: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0666 on node default medium Apr 24 23:52:16.353: INFO: Waiting up to 5m0s for pod "pod-8efd9d8f-43e7-41ee-9d62-0976233e6686" in namespace "emptydir-3229" to be "Succeeded or Failed" Apr 24 23:52:16.356: INFO: Pod "pod-8efd9d8f-43e7-41ee-9d62-0976233e6686": Phase="Pending", Reason="", readiness=false. Elapsed: 2.722844ms Apr 24 23:52:18.359: INFO: Pod "pod-8efd9d8f-43e7-41ee-9d62-0976233e6686": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006380952s Apr 24 23:52:20.363: INFO: Pod "pod-8efd9d8f-43e7-41ee-9d62-0976233e6686": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010214965s STEP: Saw pod success Apr 24 23:52:20.363: INFO: Pod "pod-8efd9d8f-43e7-41ee-9d62-0976233e6686" satisfied condition "Succeeded or Failed" Apr 24 23:52:20.366: INFO: Trying to get logs from node latest-worker pod pod-8efd9d8f-43e7-41ee-9d62-0976233e6686 container test-container: STEP: delete the pod Apr 24 23:52:20.486: INFO: Waiting for pod pod-8efd9d8f-43e7-41ee-9d62-0976233e6686 to disappear Apr 24 23:52:20.524: INFO: Pod pod-8efd9d8f-43e7-41ee-9d62-0976233e6686 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:52:20.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3229" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":66,"skipped":1154,"failed":0} S ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:52:20.531: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Apr 24 23:52:28.632: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:28.661: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:30.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:30.665: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:32.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:32.665: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:34.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:34.666: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:36.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:36.666: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:38.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:38.666: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:40.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:40.665: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:42.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:42.666: INFO: Pod pod-with-prestop-exec-hook still exists Apr 24 23:52:44.661: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 24 23:52:44.665: INFO: Pod pod-with-prestop-exec-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:52:44.673: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-4781" for this suite. • [SLOW TEST:24.151 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":275,"completed":67,"skipped":1155,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Servers with support for Table transformation /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:52:44.683: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename tables STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Servers with support for Table transformation /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/table_conversion.go:47 [It] should return a 406 for a backend which does not implement metadata [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [sig-api-machinery] Servers with support for Table transformation /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:52:44.729: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "tables-3541" for this suite. •{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":275,"completed":68,"skipped":1217,"failed":0} SSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:52:44.771: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod liveness-fe834ed1-1dbc-47cb-b59d-02cbe12a213e in namespace container-probe-8884 Apr 24 23:52:48.844: INFO: Started pod liveness-fe834ed1-1dbc-47cb-b59d-02cbe12a213e in namespace container-probe-8884 STEP: checking the pod's current state and verifying that restartCount is present Apr 24 23:52:48.848: INFO: Initial restart count of pod liveness-fe834ed1-1dbc-47cb-b59d-02cbe12a213e is 0 Apr 24 23:53:06.892: INFO: Restart count of pod container-probe-8884/liveness-fe834ed1-1dbc-47cb-b59d-02cbe12a213e is now 1 (18.044552207s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:53:06.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-8884" for this suite. • [SLOW TEST:22.174 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":275,"completed":69,"skipped":1229,"failed":0} SSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:53:06.945: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] custom resource defaulting for requests and from storage works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 24 23:53:07.029: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:53:08.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-1822" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance]","total":275,"completed":70,"skipped":1240,"failed":0} SSSSSSSSS ------------------------------ [k8s.io] Pods should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:53:08.416: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:178 [It] should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Apr 24 23:53:12.999: INFO: Successfully updated pod "pod-update-7dcea20f-acf6-420e-a90e-11f2611bfa54" STEP: verifying the updated pod is in kubernetes Apr 24 23:53:13.019: INFO: Pod update OK [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:53:13.019: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8128" for this suite. •{"msg":"PASSED [k8s.io] Pods should be updated [NodeConformance] [Conformance]","total":275,"completed":71,"skipped":1249,"failed":0} SSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:53:13.038: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod liveness-2f9ee2d7-461a-4913-88f8-03e7d65d2209 in namespace container-probe-4176 Apr 24 23:53:17.127: INFO: Started pod liveness-2f9ee2d7-461a-4913-88f8-03e7d65d2209 in namespace container-probe-4176 STEP: checking the pod's current state and verifying that restartCount is present Apr 24 23:53:17.130: INFO: Initial restart count of pod liveness-2f9ee2d7-461a-4913-88f8-03e7d65d2209 is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:57:17.732: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-4176" for this suite. • [SLOW TEST:244.718 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]","total":275,"completed":72,"skipped":1252,"failed":0} SSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:57:17.756: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [BeforeEach] when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:82 [It] should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 24 23:57:21.964: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-6098" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","total":275,"completed":73,"skipped":1266,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 24 23:57:21.974: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod busybox-4e8e3146-1300-4918-aa66-279ff0d398aa in namespace container-probe-3724 Apr 24 23:57:26.063: INFO: Started pod busybox-4e8e3146-1300-4918-aa66-279ff0d398aa in namespace container-probe-3724 STEP: checking the pod's current state and verifying that restartCount is present Apr 24 23:57:26.066: INFO: Initial restart count of pod busybox-4e8e3146-1300-4918-aa66-279ff0d398aa is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:01:26.683: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-3724" for this suite. • [SLOW TEST:244.739 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":275,"completed":74,"skipped":1291,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:01:26.713: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:01:26.798: INFO: Waiting up to 5m0s for pod "downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab" in namespace "downward-api-994" to be "Succeeded or Failed" Apr 25 00:01:26.801: INFO: Pod "downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab": Phase="Pending", Reason="", readiness=false. Elapsed: 2.688292ms Apr 25 00:01:28.804: INFO: Pod "downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006535146s Apr 25 00:01:30.809: INFO: Pod "downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011419842s STEP: Saw pod success Apr 25 00:01:30.809: INFO: Pod "downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab" satisfied condition "Succeeded or Failed" Apr 25 00:01:30.812: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab container client-container: STEP: delete the pod Apr 25 00:01:30.860: INFO: Waiting for pod downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab to disappear Apr 25 00:01:30.872: INFO: Pod downwardapi-volume-914b340a-891f-4632-9912-d9d7b65ef9ab no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:01:30.872: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-994" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":275,"completed":75,"skipped":1323,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:01:30.899: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a test externalName service STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8506.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8506.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local; sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:01:37.040: INFO: DNS probes using dns-test-d0dac4a6-b998-4b4f-bce5-7916c97a0cdd succeeded STEP: deleting the pod STEP: changing the externalName to bar.example.com STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8506.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8506.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local; sleep 1; done STEP: creating a second pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:01:43.139: INFO: File wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local from pod dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 25 00:01:43.142: INFO: File jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local from pod dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 25 00:01:43.142: INFO: Lookups using dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f failed for: [wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local] Apr 25 00:01:48.147: INFO: File wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local from pod dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 25 00:01:48.152: INFO: Lookups using dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f failed for: [wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local] Apr 25 00:01:53.147: INFO: File wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local from pod dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 25 00:01:53.151: INFO: File jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local from pod dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f contains '' instead of 'bar.example.com.' Apr 25 00:01:53.151: INFO: Lookups using dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f failed for: [wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local] Apr 25 00:01:58.151: INFO: File jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local from pod dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 25 00:01:58.151: INFO: Lookups using dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f failed for: [jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local] Apr 25 00:02:03.150: INFO: File jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local from pod dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 25 00:02:03.150: INFO: Lookups using dns-8506/dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f failed for: [jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local] Apr 25 00:02:08.149: INFO: DNS probes using dns-test-438d8b1e-f53c-474a-b87d-ccdad6f7283f succeeded STEP: deleting the pod STEP: changing the service to type=ClusterIP STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8506.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-8506.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8506.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-8506.svc.cluster.local; sleep 1; done STEP: creating a third pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:02:16.614: INFO: DNS probes using dns-test-88cfa155-fa6c-4112-aec2-64750fdedefa succeeded STEP: deleting the pod STEP: deleting the test externalName service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:02:16.691: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-8506" for this suite. • [SLOW TEST:45.802 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","total":275,"completed":76,"skipped":1346,"failed":0} SSSSSSSSS ------------------------------ [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:02:16.702: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching orphans and release non-matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: Orphaning one of the Job's Pods Apr 25 00:02:23.420: INFO: Successfully updated pod "adopt-release-48jqr" STEP: Checking that the Job readopts the Pod Apr 25 00:02:23.420: INFO: Waiting up to 15m0s for pod "adopt-release-48jqr" in namespace "job-1791" to be "adopted" Apr 25 00:02:23.426: INFO: Pod "adopt-release-48jqr": Phase="Running", Reason="", readiness=true. Elapsed: 5.26416ms Apr 25 00:02:25.430: INFO: Pod "adopt-release-48jqr": Phase="Running", Reason="", readiness=true. Elapsed: 2.009506999s Apr 25 00:02:25.430: INFO: Pod "adopt-release-48jqr" satisfied condition "adopted" STEP: Removing the labels from the Job's Pod Apr 25 00:02:25.939: INFO: Successfully updated pod "adopt-release-48jqr" STEP: Checking that the Job releases the Pod Apr 25 00:02:25.939: INFO: Waiting up to 15m0s for pod "adopt-release-48jqr" in namespace "job-1791" to be "released" Apr 25 00:02:25.946: INFO: Pod "adopt-release-48jqr": Phase="Running", Reason="", readiness=true. Elapsed: 6.274935ms Apr 25 00:02:27.950: INFO: Pod "adopt-release-48jqr": Phase="Running", Reason="", readiness=true. Elapsed: 2.010371007s Apr 25 00:02:27.950: INFO: Pod "adopt-release-48jqr" satisfied condition "released" [AfterEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:02:27.950: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-1791" for this suite. • [SLOW TEST:11.256 seconds] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching orphans and release non-matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","total":275,"completed":77,"skipped":1355,"failed":0} SSSS ------------------------------ [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:02:27.958: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:02:32.146: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-1317" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":78,"skipped":1359,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:02:32.155: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:178 [It] should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:02:36.285: INFO: Waiting up to 5m0s for pod "client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936" in namespace "pods-1497" to be "Succeeded or Failed" Apr 25 00:02:36.289: INFO: Pod "client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936": Phase="Pending", Reason="", readiness=false. Elapsed: 3.844002ms Apr 25 00:02:38.293: INFO: Pod "client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008227744s Apr 25 00:02:40.298: INFO: Pod "client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012825204s STEP: Saw pod success Apr 25 00:02:40.298: INFO: Pod "client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936" satisfied condition "Succeeded or Failed" Apr 25 00:02:40.302: INFO: Trying to get logs from node latest-worker pod client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936 container env3cont: STEP: delete the pod Apr 25 00:02:40.340: INFO: Waiting for pod client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936 to disappear Apr 25 00:02:40.389: INFO: Pod client-envvars-b2b82dca-5ff3-4d8a-861a-a68876568936 no longer exists [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:02:40.389: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-1497" for this suite. • [SLOW TEST:8.242 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":275,"completed":79,"skipped":1400,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:02:40.398: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod pod-subpath-test-configmap-2lgj STEP: Creating a pod to test atomic-volume-subpath Apr 25 00:02:40.517: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-2lgj" in namespace "subpath-2983" to be "Succeeded or Failed" Apr 25 00:02:40.521: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Pending", Reason="", readiness=false. Elapsed: 4.404775ms Apr 25 00:02:42.524: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006957855s Apr 25 00:02:44.527: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 4.01044715s Apr 25 00:02:46.532: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 6.014723713s Apr 25 00:02:48.536: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 8.019702264s Apr 25 00:02:50.541: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 10.023990546s Apr 25 00:02:52.545: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 12.028144397s Apr 25 00:02:54.550: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 14.032732653s Apr 25 00:02:56.558: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 16.04117268s Apr 25 00:02:58.562: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 18.04548674s Apr 25 00:03:00.567: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 20.049968324s Apr 25 00:03:02.571: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Running", Reason="", readiness=true. Elapsed: 22.054362302s Apr 25 00:03:05.994: INFO: Pod "pod-subpath-test-configmap-2lgj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 25.477065663s STEP: Saw pod success Apr 25 00:03:05.994: INFO: Pod "pod-subpath-test-configmap-2lgj" satisfied condition "Succeeded or Failed" Apr 25 00:03:06.043: INFO: Trying to get logs from node latest-worker pod pod-subpath-test-configmap-2lgj container test-container-subpath-configmap-2lgj: STEP: delete the pod Apr 25 00:03:06.146: INFO: Waiting for pod pod-subpath-test-configmap-2lgj to disappear Apr 25 00:03:06.150: INFO: Pod pod-subpath-test-configmap-2lgj no longer exists STEP: Deleting pod pod-subpath-test-configmap-2lgj Apr 25 00:03:06.150: INFO: Deleting pod "pod-subpath-test-configmap-2lgj" in namespace "subpath-2983" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:03:06.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-2983" for this suite. • [SLOW TEST:25.760 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]","total":275,"completed":80,"skipped":1437,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:03:06.158: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:84 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:99 STEP: Creating service test in namespace statefulset-4239 [It] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a new StatefulSet Apr 25 00:03:07.919: INFO: Found 0 stateful pods, waiting for 3 Apr 25 00:03:17.923: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:03:17.923: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:03:17.923: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Apr 25 00:03:27.923: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:03:27.923: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:03:27.923: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:03:27.953: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-4239 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:03:30.490: INFO: stderr: "I0425 00:03:30.358200 512 log.go:172] (0xc00088c6e0) (0xc0008801e0) Create stream\nI0425 00:03:30.358235 512 log.go:172] (0xc00088c6e0) (0xc0008801e0) Stream added, broadcasting: 1\nI0425 00:03:30.361392 512 log.go:172] (0xc00088c6e0) Reply frame received for 1\nI0425 00:03:30.361460 512 log.go:172] (0xc00088c6e0) (0xc0008192c0) Create stream\nI0425 00:03:30.361498 512 log.go:172] (0xc00088c6e0) (0xc0008192c0) Stream added, broadcasting: 3\nI0425 00:03:30.363099 512 log.go:172] (0xc00088c6e0) Reply frame received for 3\nI0425 00:03:30.363164 512 log.go:172] (0xc00088c6e0) (0xc000880280) Create stream\nI0425 00:03:30.363183 512 log.go:172] (0xc00088c6e0) (0xc000880280) Stream added, broadcasting: 5\nI0425 00:03:30.365663 512 log.go:172] (0xc00088c6e0) Reply frame received for 5\nI0425 00:03:30.444397 512 log.go:172] (0xc00088c6e0) Data frame received for 5\nI0425 00:03:30.444424 512 log.go:172] (0xc000880280) (5) Data frame handling\nI0425 00:03:30.444437 512 log.go:172] (0xc000880280) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:03:30.482394 512 log.go:172] (0xc00088c6e0) Data frame received for 3\nI0425 00:03:30.482420 512 log.go:172] (0xc0008192c0) (3) Data frame handling\nI0425 00:03:30.482428 512 log.go:172] (0xc0008192c0) (3) Data frame sent\nI0425 00:03:30.482541 512 log.go:172] (0xc00088c6e0) Data frame received for 3\nI0425 00:03:30.482573 512 log.go:172] (0xc0008192c0) (3) Data frame handling\nI0425 00:03:30.482728 512 log.go:172] (0xc00088c6e0) Data frame received for 5\nI0425 00:03:30.482757 512 log.go:172] (0xc000880280) (5) Data frame handling\nI0425 00:03:30.484730 512 log.go:172] (0xc00088c6e0) Data frame received for 1\nI0425 00:03:30.484745 512 log.go:172] (0xc0008801e0) (1) Data frame handling\nI0425 00:03:30.484757 512 log.go:172] (0xc0008801e0) (1) Data frame sent\nI0425 00:03:30.484786 512 log.go:172] (0xc00088c6e0) (0xc0008801e0) Stream removed, broadcasting: 1\nI0425 00:03:30.485054 512 log.go:172] (0xc00088c6e0) Go away received\nI0425 00:03:30.485203 512 log.go:172] (0xc00088c6e0) (0xc0008801e0) Stream removed, broadcasting: 1\nI0425 00:03:30.485227 512 log.go:172] (0xc00088c6e0) (0xc0008192c0) Stream removed, broadcasting: 3\nI0425 00:03:30.485236 512 log.go:172] (0xc00088c6e0) (0xc000880280) Stream removed, broadcasting: 5\n" Apr 25 00:03:30.490: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:03:30.490: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' STEP: Updating StatefulSet template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine Apr 25 00:03:40.521: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Updating Pods in reverse ordinal order Apr 25 00:03:50.599: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-4239 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:03:50.842: INFO: stderr: "I0425 00:03:50.729692 545 log.go:172] (0xc00003a790) (0xc00069f180) Create stream\nI0425 00:03:50.729750 545 log.go:172] (0xc00003a790) (0xc00069f180) Stream added, broadcasting: 1\nI0425 00:03:50.732201 545 log.go:172] (0xc00003a790) Reply frame received for 1\nI0425 00:03:50.732236 545 log.go:172] (0xc00003a790) (0xc000990000) Create stream\nI0425 00:03:50.732252 545 log.go:172] (0xc00003a790) (0xc000990000) Stream added, broadcasting: 3\nI0425 00:03:50.733097 545 log.go:172] (0xc00003a790) Reply frame received for 3\nI0425 00:03:50.733236 545 log.go:172] (0xc00003a790) (0xc00069f360) Create stream\nI0425 00:03:50.733259 545 log.go:172] (0xc00003a790) (0xc00069f360) Stream added, broadcasting: 5\nI0425 00:03:50.734178 545 log.go:172] (0xc00003a790) Reply frame received for 5\nI0425 00:03:50.833746 545 log.go:172] (0xc00003a790) Data frame received for 5\nI0425 00:03:50.833795 545 log.go:172] (0xc00069f360) (5) Data frame handling\nI0425 00:03:50.833838 545 log.go:172] (0xc00069f360) (5) Data frame sent\nI0425 00:03:50.833863 545 log.go:172] (0xc00003a790) Data frame received for 5\nI0425 00:03:50.833878 545 log.go:172] (0xc00069f360) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI0425 00:03:50.833910 545 log.go:172] (0xc00003a790) Data frame received for 3\nI0425 00:03:50.833957 545 log.go:172] (0xc000990000) (3) Data frame handling\nI0425 00:03:50.834005 545 log.go:172] (0xc000990000) (3) Data frame sent\nI0425 00:03:50.834026 545 log.go:172] (0xc00003a790) Data frame received for 3\nI0425 00:03:50.834037 545 log.go:172] (0xc000990000) (3) Data frame handling\nI0425 00:03:50.835579 545 log.go:172] (0xc00003a790) Data frame received for 1\nI0425 00:03:50.835611 545 log.go:172] (0xc00069f180) (1) Data frame handling\nI0425 00:03:50.835643 545 log.go:172] (0xc00069f180) (1) Data frame sent\nI0425 00:03:50.835671 545 log.go:172] (0xc00003a790) (0xc00069f180) Stream removed, broadcasting: 1\nI0425 00:03:50.835698 545 log.go:172] (0xc00003a790) Go away received\nI0425 00:03:50.836151 545 log.go:172] (0xc00003a790) (0xc00069f180) Stream removed, broadcasting: 1\nI0425 00:03:50.836177 545 log.go:172] (0xc00003a790) (0xc000990000) Stream removed, broadcasting: 3\nI0425 00:03:50.836190 545 log.go:172] (0xc00003a790) (0xc00069f360) Stream removed, broadcasting: 5\n" Apr 25 00:03:50.842: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:03:50.842: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:04:00.864: INFO: Waiting for StatefulSet statefulset-4239/ss2 to complete update Apr 25 00:04:00.864: INFO: Waiting for Pod statefulset-4239/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Apr 25 00:04:00.864: INFO: Waiting for Pod statefulset-4239/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Apr 25 00:04:00.864: INFO: Waiting for Pod statefulset-4239/ss2-2 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Apr 25 00:04:10.872: INFO: Waiting for StatefulSet statefulset-4239/ss2 to complete update Apr 25 00:04:10.872: INFO: Waiting for Pod statefulset-4239/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Apr 25 00:04:10.872: INFO: Waiting for Pod statefulset-4239/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Apr 25 00:04:20.884: INFO: Waiting for StatefulSet statefulset-4239/ss2 to complete update Apr 25 00:04:20.885: INFO: Waiting for Pod statefulset-4239/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 STEP: Rolling back to a previous revision Apr 25 00:04:30.873: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-4239 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:04:31.146: INFO: stderr: "I0425 00:04:31.003935 566 log.go:172] (0xc0004c0dc0) (0xc0009e40a0) Create stream\nI0425 00:04:31.003993 566 log.go:172] (0xc0004c0dc0) (0xc0009e40a0) Stream added, broadcasting: 1\nI0425 00:04:31.006415 566 log.go:172] (0xc0004c0dc0) Reply frame received for 1\nI0425 00:04:31.006455 566 log.go:172] (0xc0004c0dc0) (0xc00069f2c0) Create stream\nI0425 00:04:31.006477 566 log.go:172] (0xc0004c0dc0) (0xc00069f2c0) Stream added, broadcasting: 3\nI0425 00:04:31.007372 566 log.go:172] (0xc0004c0dc0) Reply frame received for 3\nI0425 00:04:31.007400 566 log.go:172] (0xc0004c0dc0) (0xc000aa2000) Create stream\nI0425 00:04:31.007409 566 log.go:172] (0xc0004c0dc0) (0xc000aa2000) Stream added, broadcasting: 5\nI0425 00:04:31.008422 566 log.go:172] (0xc0004c0dc0) Reply frame received for 5\nI0425 00:04:31.096592 566 log.go:172] (0xc0004c0dc0) Data frame received for 5\nI0425 00:04:31.096639 566 log.go:172] (0xc000aa2000) (5) Data frame handling\nI0425 00:04:31.096675 566 log.go:172] (0xc000aa2000) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:04:31.138225 566 log.go:172] (0xc0004c0dc0) Data frame received for 3\nI0425 00:04:31.138283 566 log.go:172] (0xc00069f2c0) (3) Data frame handling\nI0425 00:04:31.138326 566 log.go:172] (0xc00069f2c0) (3) Data frame sent\nI0425 00:04:31.138347 566 log.go:172] (0xc0004c0dc0) Data frame received for 3\nI0425 00:04:31.138358 566 log.go:172] (0xc00069f2c0) (3) Data frame handling\nI0425 00:04:31.138398 566 log.go:172] (0xc0004c0dc0) Data frame received for 5\nI0425 00:04:31.138414 566 log.go:172] (0xc000aa2000) (5) Data frame handling\nI0425 00:04:31.140317 566 log.go:172] (0xc0004c0dc0) Data frame received for 1\nI0425 00:04:31.140338 566 log.go:172] (0xc0009e40a0) (1) Data frame handling\nI0425 00:04:31.140359 566 log.go:172] (0xc0009e40a0) (1) Data frame sent\nI0425 00:04:31.140541 566 log.go:172] (0xc0004c0dc0) (0xc0009e40a0) Stream removed, broadcasting: 1\nI0425 00:04:31.141046 566 log.go:172] (0xc0004c0dc0) (0xc0009e40a0) Stream removed, broadcasting: 1\nI0425 00:04:31.141082 566 log.go:172] (0xc0004c0dc0) (0xc00069f2c0) Stream removed, broadcasting: 3\nI0425 00:04:31.141508 566 log.go:172] (0xc0004c0dc0) (0xc000aa2000) Stream removed, broadcasting: 5\n" Apr 25 00:04:31.146: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:04:31.146: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:04:41.195: INFO: Updating stateful set ss2 STEP: Rolling back update in reverse ordinal order Apr 25 00:04:51.249: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-4239 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:04:51.488: INFO: stderr: "I0425 00:04:51.388925 587 log.go:172] (0xc00003a420) (0xc0006b3220) Create stream\nI0425 00:04:51.388988 587 log.go:172] (0xc00003a420) (0xc0006b3220) Stream added, broadcasting: 1\nI0425 00:04:51.392049 587 log.go:172] (0xc00003a420) Reply frame received for 1\nI0425 00:04:51.392105 587 log.go:172] (0xc00003a420) (0xc000a3a000) Create stream\nI0425 00:04:51.392121 587 log.go:172] (0xc00003a420) (0xc000a3a000) Stream added, broadcasting: 3\nI0425 00:04:51.392988 587 log.go:172] (0xc00003a420) Reply frame received for 3\nI0425 00:04:51.393029 587 log.go:172] (0xc00003a420) (0xc000a3a0a0) Create stream\nI0425 00:04:51.393040 587 log.go:172] (0xc00003a420) (0xc000a3a0a0) Stream added, broadcasting: 5\nI0425 00:04:51.394249 587 log.go:172] (0xc00003a420) Reply frame received for 5\nI0425 00:04:51.480125 587 log.go:172] (0xc00003a420) Data frame received for 3\nI0425 00:04:51.480169 587 log.go:172] (0xc000a3a000) (3) Data frame handling\nI0425 00:04:51.480191 587 log.go:172] (0xc000a3a000) (3) Data frame sent\nI0425 00:04:51.480241 587 log.go:172] (0xc00003a420) Data frame received for 5\nI0425 00:04:51.480258 587 log.go:172] (0xc000a3a0a0) (5) Data frame handling\nI0425 00:04:51.480276 587 log.go:172] (0xc000a3a0a0) (5) Data frame sent\nI0425 00:04:51.480308 587 log.go:172] (0xc00003a420) Data frame received for 5\nI0425 00:04:51.480323 587 log.go:172] (0xc000a3a0a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI0425 00:04:51.480411 587 log.go:172] (0xc00003a420) Data frame received for 3\nI0425 00:04:51.480426 587 log.go:172] (0xc000a3a000) (3) Data frame handling\nI0425 00:04:51.482471 587 log.go:172] (0xc00003a420) Data frame received for 1\nI0425 00:04:51.482505 587 log.go:172] (0xc0006b3220) (1) Data frame handling\nI0425 00:04:51.482524 587 log.go:172] (0xc0006b3220) (1) Data frame sent\nI0425 00:04:51.482538 587 log.go:172] (0xc00003a420) (0xc0006b3220) Stream removed, broadcasting: 1\nI0425 00:04:51.482559 587 log.go:172] (0xc00003a420) Go away received\nI0425 00:04:51.483148 587 log.go:172] (0xc00003a420) (0xc0006b3220) Stream removed, broadcasting: 1\nI0425 00:04:51.483177 587 log.go:172] (0xc00003a420) (0xc000a3a000) Stream removed, broadcasting: 3\nI0425 00:04:51.483189 587 log.go:172] (0xc00003a420) (0xc000a3a0a0) Stream removed, broadcasting: 5\n" Apr 25 00:04:51.488: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:04:51.488: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:05:21.510: INFO: Waiting for StatefulSet statefulset-4239/ss2 to complete update Apr 25 00:05:21.510: INFO: Waiting for Pod statefulset-4239/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:110 Apr 25 00:05:31.533: INFO: Deleting all statefulset in ns statefulset-4239 Apr 25 00:05:31.536: INFO: Scaling statefulset ss2 to 0 Apr 25 00:06:01.561: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:06:01.563: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:06:01.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-4239" for this suite. • [SLOW TEST:175.424 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","total":275,"completed":81,"skipped":1492,"failed":0} SS ------------------------------ [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:06:01.583: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:06:01.651: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-d365bd28-6416-40c4-bc8d-d8ece986a54e" in namespace "security-context-test-1264" to be "Succeeded or Failed" Apr 25 00:06:01.657: INFO: Pod "busybox-privileged-false-d365bd28-6416-40c4-bc8d-d8ece986a54e": Phase="Pending", Reason="", readiness=false. Elapsed: 6.531538ms Apr 25 00:06:03.701: INFO: Pod "busybox-privileged-false-d365bd28-6416-40c4-bc8d-d8ece986a54e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.049966299s Apr 25 00:06:05.705: INFO: Pod "busybox-privileged-false-d365bd28-6416-40c4-bc8d-d8ece986a54e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.054153169s Apr 25 00:06:05.705: INFO: Pod "busybox-privileged-false-d365bd28-6416-40c4-bc8d-d8ece986a54e" satisfied condition "Succeeded or Failed" Apr 25 00:06:05.723: INFO: Got logs for pod "busybox-privileged-false-d365bd28-6416-40c4-bc8d-d8ece986a54e": "ip: RTNETLINK answers: Operation not permitted\n" [AfterEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:06:05.723: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-1264" for this suite. •{"msg":"PASSED [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":82,"skipped":1494,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:06:05.732: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:06:05.790: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace STEP: Creating rc "condition-test" that asks for more than the allowed pod quota STEP: Checking rc "condition-test" has the desired failure condition set STEP: Scaling down rc "condition-test" to satisfy pod quota Apr 25 00:06:08.421: INFO: Updating replication controller "condition-test" STEP: Checking rc "condition-test" has no failure condition set [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:06:08.429: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-6405" for this suite. •{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":275,"completed":83,"skipped":1518,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:06:08.490: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:06:08.629: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c" in namespace "downward-api-5054" to be "Succeeded or Failed" Apr 25 00:06:08.632: INFO: Pod "downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c": Phase="Pending", Reason="", readiness=false. Elapsed: 3.163026ms Apr 25 00:06:10.638: INFO: Pod "downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009265058s Apr 25 00:06:12.643: INFO: Pod "downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013720411s STEP: Saw pod success Apr 25 00:06:12.643: INFO: Pod "downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c" satisfied condition "Succeeded or Failed" Apr 25 00:06:12.646: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c container client-container: STEP: delete the pod Apr 25 00:06:12.699: INFO: Waiting for pod downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c to disappear Apr 25 00:06:12.715: INFO: Pod downwardapi-volume-b9800764-dbb2-4418-80a8-a190bc3aa33c no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:06:12.715: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5054" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","total":275,"completed":84,"skipped":1533,"failed":0} SSSSSSSSSS ------------------------------ [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:06:12.721: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3040 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-3040;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3040 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-3040;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3040.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-3040.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3040.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-3040.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-3040.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-3040.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3040.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 19.207.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.207.19_udp@PTR;check="$$(dig +tcp +noall +answer +search 19.207.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.207.19_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3040 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-3040;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3040 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-3040;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3040.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-3040.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3040.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-3040.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-3040.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-3040.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3040.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-3040.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3040.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 19.207.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.207.19_udp@PTR;check="$$(dig +tcp +noall +answer +search 19.207.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.207.19_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:06:18.895: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.899: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.903: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.906: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.909: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.912: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.915: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.917: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.935: INFO: Unable to read jessie_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.938: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.941: INFO: Unable to read jessie_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.944: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.947: INFO: Unable to read jessie_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.950: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.953: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.956: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:18.976: INFO: Lookups using dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-3040 wheezy_tcp@dns-test-service.dns-3040 wheezy_udp@dns-test-service.dns-3040.svc wheezy_tcp@dns-test-service.dns-3040.svc wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-3040 jessie_tcp@dns-test-service.dns-3040 jessie_udp@dns-test-service.dns-3040.svc jessie_tcp@dns-test-service.dns-3040.svc jessie_udp@_http._tcp.dns-test-service.dns-3040.svc jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc] Apr 25 00:06:23.980: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:23.984: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:23.987: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:23.990: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:23.993: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:23.996: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:23.999: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.003: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.022: INFO: Unable to read jessie_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.025: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.028: INFO: Unable to read jessie_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.031: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.034: INFO: Unable to read jessie_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.036: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.040: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.043: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:24.063: INFO: Lookups using dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-3040 wheezy_tcp@dns-test-service.dns-3040 wheezy_udp@dns-test-service.dns-3040.svc wheezy_tcp@dns-test-service.dns-3040.svc wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-3040 jessie_tcp@dns-test-service.dns-3040 jessie_udp@dns-test-service.dns-3040.svc jessie_tcp@dns-test-service.dns-3040.svc jessie_udp@_http._tcp.dns-test-service.dns-3040.svc jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc] Apr 25 00:06:28.981: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:28.984: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:28.988: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:28.991: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:28.994: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:28.996: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:28.999: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.002: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.021: INFO: Unable to read jessie_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.023: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.027: INFO: Unable to read jessie_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.030: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.033: INFO: Unable to read jessie_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.035: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.038: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.040: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:29.057: INFO: Lookups using dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-3040 wheezy_tcp@dns-test-service.dns-3040 wheezy_udp@dns-test-service.dns-3040.svc wheezy_tcp@dns-test-service.dns-3040.svc wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-3040 jessie_tcp@dns-test-service.dns-3040 jessie_udp@dns-test-service.dns-3040.svc jessie_tcp@dns-test-service.dns-3040.svc jessie_udp@_http._tcp.dns-test-service.dns-3040.svc jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc] Apr 25 00:06:33.980: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:33.985: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:33.988: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:33.992: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:33.995: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:33.998: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.002: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.005: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.022: INFO: Unable to read jessie_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.025: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.027: INFO: Unable to read jessie_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.030: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.033: INFO: Unable to read jessie_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.036: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.039: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.042: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:34.062: INFO: Lookups using dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-3040 wheezy_tcp@dns-test-service.dns-3040 wheezy_udp@dns-test-service.dns-3040.svc wheezy_tcp@dns-test-service.dns-3040.svc wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-3040 jessie_tcp@dns-test-service.dns-3040 jessie_udp@dns-test-service.dns-3040.svc jessie_tcp@dns-test-service.dns-3040.svc jessie_udp@_http._tcp.dns-test-service.dns-3040.svc jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc] Apr 25 00:06:38.981: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:38.985: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.003: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.006: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.009: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.012: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.015: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.018: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.059: INFO: Unable to read jessie_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.062: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.064: INFO: Unable to read jessie_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.067: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.070: INFO: Unable to read jessie_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.073: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.076: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.079: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:39.098: INFO: Lookups using dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-3040 wheezy_tcp@dns-test-service.dns-3040 wheezy_udp@dns-test-service.dns-3040.svc wheezy_tcp@dns-test-service.dns-3040.svc wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-3040 jessie_tcp@dns-test-service.dns-3040 jessie_udp@dns-test-service.dns-3040.svc jessie_tcp@dns-test-service.dns-3040.svc jessie_udp@_http._tcp.dns-test-service.dns-3040.svc jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc] Apr 25 00:06:43.981: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:43.985: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:43.988: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:43.991: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:43.994: INFO: Unable to read wheezy_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:43.997: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.000: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.004: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.024: INFO: Unable to read jessie_udp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.027: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.030: INFO: Unable to read jessie_udp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.033: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040 from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.036: INFO: Unable to read jessie_udp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.039: INFO: Unable to read jessie_tcp@dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.042: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.044: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc from pod dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07: the server could not find the requested resource (get pods dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07) Apr 25 00:06:44.058: INFO: Lookups using dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-3040 wheezy_tcp@dns-test-service.dns-3040 wheezy_udp@dns-test-service.dns-3040.svc wheezy_tcp@dns-test-service.dns-3040.svc wheezy_udp@_http._tcp.dns-test-service.dns-3040.svc wheezy_tcp@_http._tcp.dns-test-service.dns-3040.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-3040 jessie_tcp@dns-test-service.dns-3040 jessie_udp@dns-test-service.dns-3040.svc jessie_tcp@dns-test-service.dns-3040.svc jessie_udp@_http._tcp.dns-test-service.dns-3040.svc jessie_tcp@_http._tcp.dns-test-service.dns-3040.svc] Apr 25 00:06:49.063: INFO: DNS probes using dns-3040/dns-test-622dd3d7-20cd-490e-a845-6059f7b89d07 succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:06:49.541: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-3040" for this suite. • [SLOW TEST:36.842 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","total":275,"completed":85,"skipped":1543,"failed":0} SSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:06:49.564: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test substitution in container's command Apr 25 00:06:49.714: INFO: Waiting up to 5m0s for pod "var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37" in namespace "var-expansion-3108" to be "Succeeded or Failed" Apr 25 00:06:49.719: INFO: Pod "var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37": Phase="Pending", Reason="", readiness=false. Elapsed: 4.05886ms Apr 25 00:06:51.722: INFO: Pod "var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00796551s Apr 25 00:06:53.726: INFO: Pod "var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011586053s STEP: Saw pod success Apr 25 00:06:53.726: INFO: Pod "var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37" satisfied condition "Succeeded or Failed" Apr 25 00:06:53.729: INFO: Trying to get logs from node latest-worker pod var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37 container dapi-container: STEP: delete the pod Apr 25 00:06:53.765: INFO: Waiting for pod var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37 to disappear Apr 25 00:06:53.785: INFO: Pod var-expansion-f0914e69-1c2c-4559-90e6-5c7ce4fe2c37 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:06:53.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-3108" for this suite. •{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","total":275,"completed":86,"skipped":1548,"failed":0} ------------------------------ [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:06:53.794: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:06:53.905: INFO: Waiting up to 5m0s for pod "downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577" in namespace "projected-3852" to be "Succeeded or Failed" Apr 25 00:06:53.911: INFO: Pod "downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577": Phase="Pending", Reason="", readiness=false. Elapsed: 5.292409ms Apr 25 00:06:55.935: INFO: Pod "downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029902145s Apr 25 00:06:57.939: INFO: Pod "downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.033668574s STEP: Saw pod success Apr 25 00:06:57.939: INFO: Pod "downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577" satisfied condition "Succeeded or Failed" Apr 25 00:06:57.945: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577 container client-container: STEP: delete the pod Apr 25 00:06:58.017: INFO: Waiting for pod downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577 to disappear Apr 25 00:06:58.025: INFO: Pod downwardapi-volume-838f6aef-9eb3-4ddf-bb59-f972bc34a577 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:06:58.025: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3852" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":87,"skipped":1548,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:06:58.036: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name projected-configmap-test-volume-map-c9fb0196-0b08-4b47-9ff3-daea92f27720 STEP: Creating a pod to test consume configMaps Apr 25 00:06:58.115: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc" in namespace "projected-321" to be "Succeeded or Failed" Apr 25 00:06:58.129: INFO: Pod "pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc": Phase="Pending", Reason="", readiness=false. Elapsed: 13.955749ms Apr 25 00:07:00.145: INFO: Pod "pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030352927s Apr 25 00:07:02.150: INFO: Pod "pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034492835s STEP: Saw pod success Apr 25 00:07:02.150: INFO: Pod "pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc" satisfied condition "Succeeded or Failed" Apr 25 00:07:02.153: INFO: Trying to get logs from node latest-worker pod pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc container projected-configmap-volume-test: STEP: delete the pod Apr 25 00:07:02.168: INFO: Waiting for pod pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc to disappear Apr 25 00:07:02.179: INFO: Pod pod-projected-configmaps-4707e645-fd55-4b03-af9d-0585571070dc no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:02.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-321" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":275,"completed":88,"skipped":1577,"failed":0} SS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:02.186: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:91 Apr 25 00:07:02.276: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 25 00:07:02.314: INFO: Waiting for terminating namespaces to be deleted... Apr 25 00:07:02.317: INFO: Logging pods the kubelet thinks is on node latest-worker before test Apr 25 00:07:02.346: INFO: kindnet-vnjgh from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:07:02.346: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:07:02.346: INFO: kube-proxy-s9v6p from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:07:02.346: INFO: Container kube-proxy ready: true, restart count 0 Apr 25 00:07:02.346: INFO: Logging pods the kubelet thinks is on node latest-worker2 before test Apr 25 00:07:02.351: INFO: kindnet-zq6gp from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:07:02.351: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:07:02.351: INFO: kube-proxy-c5xlk from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:07:02.351: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-9a5fdf79-16e8-477f-840a-aa04104200ec 42 STEP: Trying to relaunch the pod, now with labels. STEP: removing the label kubernetes.io/e2e-9a5fdf79-16e8-477f-840a-aa04104200ec off the node latest-worker2 STEP: verifying the node doesn't have the label kubernetes.io/e2e-9a5fdf79-16e8-477f-840a-aa04104200ec [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:10.591: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-795" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:82 • [SLOW TEST:8.412 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance]","total":275,"completed":89,"skipped":1579,"failed":0} SSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:10.599: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:07:10.656: INFO: Waiting up to 5m0s for pod "downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e" in namespace "projected-5352" to be "Succeeded or Failed" Apr 25 00:07:10.671: INFO: Pod "downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e": Phase="Pending", Reason="", readiness=false. Elapsed: 15.020559ms Apr 25 00:07:12.676: INFO: Pod "downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019737816s Apr 25 00:07:14.680: INFO: Pod "downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e": Phase="Pending", Reason="", readiness=false. Elapsed: 4.023911212s Apr 25 00:07:16.683: INFO: Pod "downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.027259185s STEP: Saw pod success Apr 25 00:07:16.683: INFO: Pod "downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e" satisfied condition "Succeeded or Failed" Apr 25 00:07:16.686: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e container client-container: STEP: delete the pod Apr 25 00:07:16.714: INFO: Waiting for pod downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e to disappear Apr 25 00:07:16.720: INFO: Pod downwardapi-volume-cf3688df-1dec-4744-9427-c230e8ebf46e no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:16.720: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5352" for this suite. • [SLOW TEST:6.128 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":90,"skipped":1588,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:16.727: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir volume type on node default medium Apr 25 00:07:16.802: INFO: Waiting up to 5m0s for pod "pod-4b908635-67c7-4884-8f6e-c570703e9592" in namespace "emptydir-4130" to be "Succeeded or Failed" Apr 25 00:07:16.810: INFO: Pod "pod-4b908635-67c7-4884-8f6e-c570703e9592": Phase="Pending", Reason="", readiness=false. Elapsed: 7.310594ms Apr 25 00:07:18.831: INFO: Pod "pod-4b908635-67c7-4884-8f6e-c570703e9592": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028489978s Apr 25 00:07:20.835: INFO: Pod "pod-4b908635-67c7-4884-8f6e-c570703e9592": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.032879566s STEP: Saw pod success Apr 25 00:07:20.835: INFO: Pod "pod-4b908635-67c7-4884-8f6e-c570703e9592" satisfied condition "Succeeded or Failed" Apr 25 00:07:20.838: INFO: Trying to get logs from node latest-worker2 pod pod-4b908635-67c7-4884-8f6e-c570703e9592 container test-container: STEP: delete the pod Apr 25 00:07:20.869: INFO: Waiting for pod pod-4b908635-67c7-4884-8f6e-c570703e9592 to disappear Apr 25 00:07:20.874: INFO: Pod pod-4b908635-67c7-4884-8f6e-c570703e9592 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:20.874: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4130" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":91,"skipped":1620,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:20.882: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name projected-configmap-test-volume-6aed6783-7e14-43f5-afc4-6a360464d51f STEP: Creating a pod to test consume configMaps Apr 25 00:07:20.964: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406" in namespace "projected-8605" to be "Succeeded or Failed" Apr 25 00:07:20.967: INFO: Pod "pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406": Phase="Pending", Reason="", readiness=false. Elapsed: 2.841831ms Apr 25 00:07:22.970: INFO: Pod "pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006524279s Apr 25 00:07:25.574: INFO: Pod "pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.609939036s STEP: Saw pod success Apr 25 00:07:25.574: INFO: Pod "pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406" satisfied condition "Succeeded or Failed" Apr 25 00:07:25.577: INFO: Trying to get logs from node latest-worker2 pod pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406 container projected-configmap-volume-test: STEP: delete the pod Apr 25 00:07:25.723: INFO: Waiting for pod pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406 to disappear Apr 25 00:07:25.773: INFO: Pod pod-projected-configmaps-990b560d-f0df-4f93-8acb-8a7d9e194406 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:25.773: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8605" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":275,"completed":92,"skipped":1635,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:25.783: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap configmap-8895/configmap-test-9f276187-f124-4498-ba45-66ce582d37e7 STEP: Creating a pod to test consume configMaps Apr 25 00:07:25.882: INFO: Waiting up to 5m0s for pod "pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d" in namespace "configmap-8895" to be "Succeeded or Failed" Apr 25 00:07:25.886: INFO: Pod "pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.129333ms Apr 25 00:07:27.889: INFO: Pod "pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007459283s Apr 25 00:07:29.894: INFO: Pod "pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012017839s STEP: Saw pod success Apr 25 00:07:29.894: INFO: Pod "pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d" satisfied condition "Succeeded or Failed" Apr 25 00:07:29.897: INFO: Trying to get logs from node latest-worker2 pod pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d container env-test: STEP: delete the pod Apr 25 00:07:29.936: INFO: Waiting for pod pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d to disappear Apr 25 00:07:29.986: INFO: Pod pod-configmaps-7f13b74a-36cc-49f8-ae36-3d774ba2257d no longer exists [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:29.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8895" for this suite. •{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","total":275,"completed":93,"skipped":1649,"failed":0} SS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:29.994: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] removes definition from spec when one version gets changed to not be served [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: set up a multi version CRD Apr 25 00:07:30.062: INFO: >>> kubeConfig: /root/.kube/config STEP: mark a version not serverd STEP: check the unserved version gets removed STEP: check the other version is not changed [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:44.309: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-4035" for this suite. • [SLOW TEST:14.322 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 removes definition from spec when one version gets changed to not be served [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":275,"completed":94,"skipped":1651,"failed":0} S ------------------------------ [k8s.io] Lease lease API should be available [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Lease /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:44.316: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename lease-test STEP: Waiting for a default service account to be provisioned in namespace [It] lease API should be available [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Lease /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:44.472: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "lease-test-3640" for this suite. •{"msg":"PASSED [k8s.io] Lease lease API should be available [Conformance]","total":275,"completed":95,"skipped":1652,"failed":0} SSSSSSSS ------------------------------ [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:44.481: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Given a Pod with a 'name' label pod-adoption-release is created STEP: When a replicaset with a matching selector is created STEP: Then the orphan pod is adopted STEP: When the matched label of one of its pods change Apr 25 00:07:49.642: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:49.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-7604" for this suite. • [SLOW TEST:5.223 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","total":275,"completed":96,"skipped":1660,"failed":0} SSSSS ------------------------------ [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:49.704: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:126 STEP: Setting up server cert STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication STEP: Deploying the custom resource conversion webhook pod STEP: Wait for the deployment to be ready Apr 25 00:07:50.210: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set Apr 25 00:07:52.285: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370070, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370070, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370070, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370070, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-54c8b67c75\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:07:55.342: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 [It] should be able to convert from CR v1 to CR v2 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:07:55.347: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating a v1 custom resource STEP: v2 custom resource should be converted [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:07:56.476: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-webhook-1811" for this suite. [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:137 • [SLOW TEST:6.874 seconds] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to convert from CR v1 to CR v2 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","total":275,"completed":97,"skipped":1665,"failed":0} S ------------------------------ [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:07:56.577: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name s-test-opt-del-2e3bf01e-9c90-41ba-b4ba-883db6857a22 STEP: Creating secret with name s-test-opt-upd-f8f51cb5-a11e-4a38-a078-15218b32ac37 STEP: Creating the pod STEP: Deleting secret s-test-opt-del-2e3bf01e-9c90-41ba-b4ba-883db6857a22 STEP: Updating secret s-test-opt-upd-f8f51cb5-a11e-4a38-a078-15218b32ac37 STEP: Creating secret with name s-test-opt-create-636f26f8-b0c3-4c40-b659-be26c5fc8978 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:09:31.195: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1345" for this suite. • [SLOW TEST:94.625 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:35 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","total":275,"completed":98,"skipped":1666,"failed":0} SSS ------------------------------ [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:09:31.202: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap configmap-4/configmap-test-ffed1898-669b-4601-bb62-7629fc396624 STEP: Creating a pod to test consume configMaps Apr 25 00:09:31.294: INFO: Waiting up to 5m0s for pod "pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc" in namespace "configmap-4" to be "Succeeded or Failed" Apr 25 00:09:31.298: INFO: Pod "pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc": Phase="Pending", Reason="", readiness=false. Elapsed: 3.550104ms Apr 25 00:09:33.301: INFO: Pod "pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006964496s Apr 25 00:09:35.306: INFO: Pod "pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011385901s STEP: Saw pod success Apr 25 00:09:35.306: INFO: Pod "pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc" satisfied condition "Succeeded or Failed" Apr 25 00:09:35.309: INFO: Trying to get logs from node latest-worker pod pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc container env-test: STEP: delete the pod Apr 25 00:09:35.341: INFO: Waiting for pod pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc to disappear Apr 25 00:09:35.346: INFO: Pod pod-configmaps-b0efb973-8d89-48b6-8614-76c747d90edc no longer exists [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:09:35.346: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4" for this suite. •{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","total":275,"completed":99,"skipped":1669,"failed":0} S ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:09:35.352: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:09:36.017: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:09:38.266: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370176, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370176, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370176, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370176, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:09:41.324: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny attaching pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering the webhook via the AdmissionRegistration API STEP: create a pod STEP: 'kubectl attach' the pod, should be denied by the webhook Apr 25 00:09:45.393: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config attach --namespace=webhook-2497 to-be-attached-pod -i -c=container1' Apr 25 00:09:45.523: INFO: rc: 1 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:09:45.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2497" for this suite. STEP: Destroying namespace "webhook-2497-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:10.280 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny attaching pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","total":275,"completed":100,"skipped":1670,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:09:45.632: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-7be35f5c-c23f-4ef6-94d3-134a2264162b STEP: Creating a pod to test consume secrets Apr 25 00:09:45.695: INFO: Waiting up to 5m0s for pod "pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51" in namespace "secrets-3902" to be "Succeeded or Failed" Apr 25 00:09:45.712: INFO: Pod "pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51": Phase="Pending", Reason="", readiness=false. Elapsed: 16.828597ms Apr 25 00:09:47.766: INFO: Pod "pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51": Phase="Pending", Reason="", readiness=false. Elapsed: 2.070943932s Apr 25 00:09:49.771: INFO: Pod "pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.075125142s STEP: Saw pod success Apr 25 00:09:49.771: INFO: Pod "pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51" satisfied condition "Succeeded or Failed" Apr 25 00:09:49.774: INFO: Trying to get logs from node latest-worker2 pod pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51 container secret-volume-test: STEP: delete the pod Apr 25 00:09:49.833: INFO: Waiting for pod pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51 to disappear Apr 25 00:09:49.843: INFO: Pod pod-secrets-8e53dbf2-92c7-41fb-b28d-12df5710db51 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:09:49.844: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3902" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","total":275,"completed":101,"skipped":1681,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:09:49.851: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating projection with secret that has name projected-secret-test-7e51b1ce-17b0-4a80-b96b-da94f9bf7449 STEP: Creating a pod to test consume secrets Apr 25 00:09:49.930: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7" in namespace "projected-7592" to be "Succeeded or Failed" Apr 25 00:09:49.952: INFO: Pod "pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7": Phase="Pending", Reason="", readiness=false. Elapsed: 21.742552ms Apr 25 00:09:51.982: INFO: Pod "pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.05200935s Apr 25 00:09:53.986: INFO: Pod "pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.056119985s STEP: Saw pod success Apr 25 00:09:53.986: INFO: Pod "pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7" satisfied condition "Succeeded or Failed" Apr 25 00:09:54.012: INFO: Trying to get logs from node latest-worker pod pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7 container projected-secret-volume-test: STEP: delete the pod Apr 25 00:09:54.028: INFO: Waiting for pod pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7 to disappear Apr 25 00:09:54.058: INFO: Pod pod-projected-secrets-fb4c52a6-fdfa-4030-a36d-782c3e8bbce7 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:09:54.058: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7592" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":102,"skipped":1747,"failed":0} S ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:09:54.066: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-volume-54b99c1d-783f-4dac-b1be-689a8f9193d4 STEP: Creating a pod to test consume configMaps Apr 25 00:09:54.137: INFO: Waiting up to 5m0s for pod "pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9" in namespace "configmap-6846" to be "Succeeded or Failed" Apr 25 00:09:54.141: INFO: Pod "pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.488641ms Apr 25 00:09:56.267: INFO: Pod "pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.130362766s Apr 25 00:09:58.271: INFO: Pod "pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9": Phase="Running", Reason="", readiness=true. Elapsed: 4.134447818s Apr 25 00:10:00.276: INFO: Pod "pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.138803456s STEP: Saw pod success Apr 25 00:10:00.276: INFO: Pod "pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9" satisfied condition "Succeeded or Failed" Apr 25 00:10:00.279: INFO: Trying to get logs from node latest-worker2 pod pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9 container configmap-volume-test: STEP: delete the pod Apr 25 00:10:00.298: INFO: Waiting for pod pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9 to disappear Apr 25 00:10:00.302: INFO: Pod pod-configmaps-b91a37a4-4c70-475b-86e8-6709bd7a5ed9 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:10:00.302: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-6846" for this suite. • [SLOW TEST:6.242 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":275,"completed":103,"skipped":1748,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:10:00.309: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Apr 25 00:10:00.419: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:00.424: INFO: Number of nodes with available pods: 0 Apr 25 00:10:00.424: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:10:01.429: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:01.433: INFO: Number of nodes with available pods: 0 Apr 25 00:10:01.433: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:10:02.432: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:02.455: INFO: Number of nodes with available pods: 0 Apr 25 00:10:02.455: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:10:03.430: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:03.433: INFO: Number of nodes with available pods: 1 Apr 25 00:10:03.433: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:04.430: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:04.437: INFO: Number of nodes with available pods: 1 Apr 25 00:10:04.437: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:05.428: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:05.430: INFO: Number of nodes with available pods: 2 Apr 25 00:10:05.430: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Stop a daemon pod, check that the daemon pod is revived. Apr 25 00:10:05.443: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:05.449: INFO: Number of nodes with available pods: 1 Apr 25 00:10:05.449: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:06.498: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:06.502: INFO: Number of nodes with available pods: 1 Apr 25 00:10:06.502: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:07.454: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:07.456: INFO: Number of nodes with available pods: 1 Apr 25 00:10:07.456: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:08.454: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:08.461: INFO: Number of nodes with available pods: 1 Apr 25 00:10:08.461: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:09.454: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:09.458: INFO: Number of nodes with available pods: 1 Apr 25 00:10:09.458: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:10.455: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:10.459: INFO: Number of nodes with available pods: 1 Apr 25 00:10:10.459: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:11.455: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:11.459: INFO: Number of nodes with available pods: 1 Apr 25 00:10:11.459: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:12.454: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:12.458: INFO: Number of nodes with available pods: 1 Apr 25 00:10:12.458: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:13.455: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:13.459: INFO: Number of nodes with available pods: 1 Apr 25 00:10:13.459: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:14.454: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:14.458: INFO: Number of nodes with available pods: 1 Apr 25 00:10:14.458: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:15.454: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:15.458: INFO: Number of nodes with available pods: 1 Apr 25 00:10:15.458: INFO: Node latest-worker2 is running more than one daemon pod Apr 25 00:10:16.454: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:10:16.458: INFO: Number of nodes with available pods: 2 Apr 25 00:10:16.458: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-8458, will wait for the garbage collector to delete the pods Apr 25 00:10:16.520: INFO: Deleting DaemonSet.extensions daemon-set took: 5.943739ms Apr 25 00:10:16.820: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.289892ms Apr 25 00:10:19.722: INFO: Number of nodes with available pods: 0 Apr 25 00:10:19.722: INFO: Number of running nodes: 0, number of available pods: 0 Apr 25 00:10:19.727: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-8458/daemonsets","resourceVersion":"10787131"},"items":null} Apr 25 00:10:19.730: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-8458/pods","resourceVersion":"10787131"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:10:19.738: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-8458" for this suite. • [SLOW TEST:19.434 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","total":275,"completed":104,"skipped":1791,"failed":0} SS ------------------------------ [sig-network] DNS should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:10:19.744: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-427.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-427.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:10:25.909: INFO: DNS probes using dns-427/dns-test-32293d71-70d6-4a93-bb9c-990000e5a3d1 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:10:25.972: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-427" for this suite. • [SLOW TEST:6.244 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for the cluster [Conformance]","total":275,"completed":105,"skipped":1793,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:10:25.989: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-volume-e8c83c4a-3abc-43a5-ad8f-cf525bdebd84 STEP: Creating a pod to test consume configMaps Apr 25 00:10:26.061: INFO: Waiting up to 5m0s for pod "pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15" in namespace "configmap-6272" to be "Succeeded or Failed" Apr 25 00:10:26.090: INFO: Pod "pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15": Phase="Pending", Reason="", readiness=false. Elapsed: 29.025032ms Apr 25 00:10:28.098: INFO: Pod "pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037042406s Apr 25 00:10:30.103: INFO: Pod "pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15": Phase="Running", Reason="", readiness=true. Elapsed: 4.041901749s Apr 25 00:10:32.108: INFO: Pod "pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.046633273s STEP: Saw pod success Apr 25 00:10:32.108: INFO: Pod "pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15" satisfied condition "Succeeded or Failed" Apr 25 00:10:32.111: INFO: Trying to get logs from node latest-worker pod pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15 container configmap-volume-test: STEP: delete the pod Apr 25 00:10:32.152: INFO: Waiting for pod pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15 to disappear Apr 25 00:10:32.199: INFO: Pod pod-configmaps-5555ee9a-f6d9-40ca-a787-679c6b3fda15 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:10:32.199: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-6272" for this suite. • [SLOW TEST:6.218 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":275,"completed":106,"skipped":1827,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:10:32.208: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod busybox-7ee7a490-cf48-4889-9ad9-258f5550e8b1 in namespace container-probe-7148 Apr 25 00:10:36.309: INFO: Started pod busybox-7ee7a490-cf48-4889-9ad9-258f5550e8b1 in namespace container-probe-7148 STEP: checking the pod's current state and verifying that restartCount is present Apr 25 00:10:36.312: INFO: Initial restart count of pod busybox-7ee7a490-cf48-4889-9ad9-258f5550e8b1 is 0 Apr 25 00:11:30.423: INFO: Restart count of pod container-probe-7148/busybox-7ee7a490-cf48-4889-9ad9-258f5550e8b1 is now 1 (54.110925069s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:11:30.449: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-7148" for this suite. • [SLOW TEST:58.272 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":275,"completed":107,"skipped":1862,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:11:30.480: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:11:31.041: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:11:33.050: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370291, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370291, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370291, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370291, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:11:36.089: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should honor timeout [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Setting timeout (1s) shorter than webhook latency (5s) STEP: Registering slow webhook via the AdmissionRegistration API STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s) STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore STEP: Registering slow webhook via the AdmissionRegistration API STEP: Having no error when timeout is longer than webhook latency STEP: Registering slow webhook via the AdmissionRegistration API STEP: Having no error when timeout is empty (defaulted to 10s in v1) STEP: Registering slow webhook via the AdmissionRegistration API [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:11:48.273: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-641" for this suite. STEP: Destroying namespace "webhook-641-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:17.913 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should honor timeout [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","total":275,"completed":108,"skipped":1884,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:11:48.393: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD with validation schema [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:11:48.442: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with known and required properties Apr 25 00:11:51.370: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 create -f -' Apr 25 00:11:54.250: INFO: stderr: "" Apr 25 00:11:54.250: INFO: stdout: "e2e-test-crd-publish-openapi-7768-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" Apr 25 00:11:54.250: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 delete e2e-test-crd-publish-openapi-7768-crds test-foo' Apr 25 00:11:54.351: INFO: stderr: "" Apr 25 00:11:54.351: INFO: stdout: "e2e-test-crd-publish-openapi-7768-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" Apr 25 00:11:54.352: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 apply -f -' Apr 25 00:11:54.609: INFO: stderr: "" Apr 25 00:11:54.609: INFO: stdout: "e2e-test-crd-publish-openapi-7768-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" Apr 25 00:11:54.609: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 delete e2e-test-crd-publish-openapi-7768-crds test-foo' Apr 25 00:11:54.722: INFO: stderr: "" Apr 25 00:11:54.722: INFO: stdout: "e2e-test-crd-publish-openapi-7768-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" STEP: client-side validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema Apr 25 00:11:54.722: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 create -f -' Apr 25 00:11:54.945: INFO: rc: 1 Apr 25 00:11:54.945: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 apply -f -' Apr 25 00:11:55.165: INFO: rc: 1 STEP: client-side validation (kubectl create and apply) rejects request without required properties Apr 25 00:11:55.165: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 create -f -' Apr 25 00:11:55.406: INFO: rc: 1 Apr 25 00:11:55.406: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-1821 apply -f -' Apr 25 00:11:55.642: INFO: rc: 1 STEP: kubectl explain works to explain CR properties Apr 25 00:11:55.642: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-7768-crds' Apr 25 00:11:55.877: INFO: stderr: "" Apr 25 00:11:55.877: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7768-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" STEP: kubectl explain works to explain CR properties recursively Apr 25 00:11:55.877: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-7768-crds.metadata' Apr 25 00:11:56.116: INFO: stderr: "" Apr 25 00:11:56.116: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7768-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n clusterName\t\n The name of the cluster which the object belongs to. This is used to\n distinguish resources with same name and namespace in different clusters.\n This field is not set anywhere right now and apiserver is going to ignore\n it if set in create or update request.\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC. Populated by the system.\n Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested. Populated by the system when a graceful deletion is\n requested. Read-only. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server. If this field is specified and the generated name exists, the\n server will NOT return a 409 - instead, it will either return 201 Created\n or 500 with Reason ServerTimeout indicating a unique name could not be\n found in the time allotted, and the client should retry (optionally after\n the time indicated in the Retry-After header). Applied only if Name is not\n specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty. Must\n be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources. Populated by the system.\n Read-only. Value must be treated as opaque by clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n SelfLink is a URL representing this object. Populated by the system.\n Read-only. DEPRECATED Kubernetes will stop propagating this field in 1.20\n release and the field is planned to be removed in 1.21 release.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations. Populated by the system. Read-only.\n More info: http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" Apr 25 00:11:56.117: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-7768-crds.spec' Apr 25 00:11:56.381: INFO: stderr: "" Apr 25 00:11:56.381: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7768-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" Apr 25 00:11:56.381: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-7768-crds.spec.bars' Apr 25 00:11:56.645: INFO: stderr: "" Apr 25 00:11:56.645: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7768-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n name\t -required-\n Name of Bar.\n\n" STEP: kubectl explain works to return error when explain is called on property that doesn't exist Apr 25 00:11:56.645: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-7768-crds.spec.bars2' Apr 25 00:11:56.886: INFO: rc: 1 [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:11:59.792: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-1821" for this suite. • [SLOW TEST:11.408 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD with validation schema [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":275,"completed":109,"skipped":1905,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:11:59.802: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [It] should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:12:03.925: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-1619" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","total":275,"completed":110,"skipped":1951,"failed":0} ------------------------------ [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:12:03.933: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [BeforeEach] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1206 STEP: creating the pod Apr 25 00:12:03.980: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-4679' Apr 25 00:12:04.276: INFO: stderr: "" Apr 25 00:12:04.276: INFO: stdout: "pod/pause created\n" Apr 25 00:12:04.276: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] Apr 25 00:12:04.276: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-4679" to be "running and ready" Apr 25 00:12:04.281: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 5.207085ms Apr 25 00:12:06.285: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008971951s Apr 25 00:12:08.289: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.013182525s Apr 25 00:12:08.289: INFO: Pod "pause" satisfied condition "running and ready" Apr 25 00:12:08.289: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] [It] should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: adding the label testing-label with value testing-label-value to a pod Apr 25 00:12:08.289: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config label pods pause testing-label=testing-label-value --namespace=kubectl-4679' Apr 25 00:12:08.381: INFO: stderr: "" Apr 25 00:12:08.382: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod has the label testing-label with the value testing-label-value Apr 25 00:12:08.382: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-4679' Apr 25 00:12:08.473: INFO: stderr: "" Apr 25 00:12:08.473: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s testing-label-value\n" STEP: removing the label testing-label of a pod Apr 25 00:12:08.473: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config label pods pause testing-label- --namespace=kubectl-4679' Apr 25 00:12:08.592: INFO: stderr: "" Apr 25 00:12:08.592: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod doesn't have the label testing-label Apr 25 00:12:08.593: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-4679' Apr 25 00:12:08.682: INFO: stderr: "" Apr 25 00:12:08.682: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s \n" [AfterEach] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1213 STEP: using delete to clean up resources Apr 25 00:12:08.682: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-4679' Apr 25 00:12:08.786: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 25 00:12:08.786: INFO: stdout: "pod \"pause\" force deleted\n" Apr 25 00:12:08.786: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get rc,svc -l name=pause --no-headers --namespace=kubectl-4679' Apr 25 00:12:08.882: INFO: stderr: "No resources found in kubectl-4679 namespace.\n" Apr 25 00:12:08.882: INFO: stdout: "" Apr 25 00:12:08.882: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -l name=pause --namespace=kubectl-4679 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 25 00:12:09.011: INFO: stderr: "" Apr 25 00:12:09.011: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:12:09.011: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4679" for this suite. • [SLOW TEST:5.148 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1203 should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance]","total":275,"completed":111,"skipped":1951,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:12:09.081: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating service multi-endpoint-test in namespace services-7588 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7588 to expose endpoints map[] Apr 25 00:12:09.284: INFO: successfully validated that service multi-endpoint-test in namespace services-7588 exposes endpoints map[] (12.110625ms elapsed) STEP: Creating pod pod1 in namespace services-7588 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7588 to expose endpoints map[pod1:[100]] Apr 25 00:12:13.383: INFO: successfully validated that service multi-endpoint-test in namespace services-7588 exposes endpoints map[pod1:[100]] (4.092684009s elapsed) STEP: Creating pod pod2 in namespace services-7588 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7588 to expose endpoints map[pod1:[100] pod2:[101]] Apr 25 00:12:16.487: INFO: successfully validated that service multi-endpoint-test in namespace services-7588 exposes endpoints map[pod1:[100] pod2:[101]] (3.09909324s elapsed) STEP: Deleting pod pod1 in namespace services-7588 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7588 to expose endpoints map[pod2:[101]] Apr 25 00:12:17.548: INFO: successfully validated that service multi-endpoint-test in namespace services-7588 exposes endpoints map[pod2:[101]] (1.05510008s elapsed) STEP: Deleting pod pod2 in namespace services-7588 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-7588 to expose endpoints map[] Apr 25 00:12:18.561: INFO: successfully validated that service multi-endpoint-test in namespace services-7588 exposes endpoints map[] (1.008921904s elapsed) [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:12:18.646: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-7588" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 • [SLOW TEST:9.584 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods [Conformance]","total":275,"completed":112,"skipped":1970,"failed":0} SSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:12:18.665: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name projected-configmap-test-volume-map-efbc8b7e-9012-44a5-a965-1db90c219ae6 STEP: Creating a pod to test consume configMaps Apr 25 00:12:18.775: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156" in namespace "projected-714" to be "Succeeded or Failed" Apr 25 00:12:18.796: INFO: Pod "pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156": Phase="Pending", Reason="", readiness=false. Elapsed: 20.740773ms Apr 25 00:12:20.829: INFO: Pod "pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156": Phase="Pending", Reason="", readiness=false. Elapsed: 2.053845198s Apr 25 00:12:22.833: INFO: Pod "pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156": Phase="Running", Reason="", readiness=true. Elapsed: 4.058510418s Apr 25 00:12:24.837: INFO: Pod "pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.062474854s STEP: Saw pod success Apr 25 00:12:24.837: INFO: Pod "pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156" satisfied condition "Succeeded or Failed" Apr 25 00:12:24.841: INFO: Trying to get logs from node latest-worker pod pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156 container projected-configmap-volume-test: STEP: delete the pod Apr 25 00:12:24.898: INFO: Waiting for pod pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156 to disappear Apr 25 00:12:24.904: INFO: Pod pod-projected-configmaps-372c11b1-f7a2-496f-9190-018e3b91e156 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:12:24.904: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-714" for this suite. • [SLOW TEST:6.245 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:36 should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":113,"skipped":1975,"failed":0} SS ------------------------------ [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:12:24.911: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-93cccb39-483a-41dd-bcde-26799db9d7b2 STEP: Creating a pod to test consume secrets Apr 25 00:12:25.038: INFO: Waiting up to 5m0s for pod "pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b" in namespace "secrets-1091" to be "Succeeded or Failed" Apr 25 00:12:25.065: INFO: Pod "pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b": Phase="Pending", Reason="", readiness=false. Elapsed: 27.132604ms Apr 25 00:12:27.070: INFO: Pod "pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031538873s Apr 25 00:12:29.074: INFO: Pod "pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035932499s STEP: Saw pod success Apr 25 00:12:29.074: INFO: Pod "pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b" satisfied condition "Succeeded or Failed" Apr 25 00:12:29.077: INFO: Trying to get logs from node latest-worker2 pod pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b container secret-volume-test: STEP: delete the pod Apr 25 00:12:29.112: INFO: Waiting for pod pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b to disappear Apr 25 00:12:29.123: INFO: Pod pod-secrets-c2ebefb7-fd63-4772-8d41-0a1099f20a6b no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:12:29.123: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-1091" for this suite. STEP: Destroying namespace "secret-namespace-3962" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":275,"completed":114,"skipped":1977,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:12:29.138: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted Apr 25 00:12:36.050: INFO: 0 pods remaining Apr 25 00:12:36.050: INFO: 0 pods has nil DeletionTimestamp Apr 25 00:12:36.050: INFO: Apr 25 00:12:36.432: INFO: 0 pods remaining Apr 25 00:12:36.432: INFO: 0 pods has nil DeletionTimestamp Apr 25 00:12:36.432: INFO: STEP: Gathering metrics W0425 00:12:37.770124 8 metrics_grabber.go:84] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 25 00:12:37.770: INFO: For apiserver_request_total: For apiserver_request_latency_seconds: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:12:37.770: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-7856" for this suite. • [SLOW TEST:8.708 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","total":275,"completed":115,"skipped":2058,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:12:37.846: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should be able to change the type from ClusterIP to ExternalName [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-4176 STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service STEP: creating service externalsvc in namespace services-4176 STEP: creating replication controller externalsvc in namespace services-4176 I0425 00:12:40.853526 8 runners.go:190] Created replication controller with name: externalsvc, namespace: services-4176, replica count: 2 I0425 00:12:43.904067 8 runners.go:190] externalsvc Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0425 00:12:46.904367 8 runners.go:190] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady STEP: changing the ClusterIP service to type=ExternalName Apr 25 00:12:46.954: INFO: Creating new exec pod Apr 25 00:12:50.975: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-4176 execpodldzr5 -- /bin/sh -x -c nslookup clusterip-service' Apr 25 00:12:51.206: INFO: stderr: "I0425 00:12:51.102616 1088 log.go:172] (0xc0009a3080) (0xc0005a65a0) Create stream\nI0425 00:12:51.102687 1088 log.go:172] (0xc0009a3080) (0xc0005a65a0) Stream added, broadcasting: 1\nI0425 00:12:51.107503 1088 log.go:172] (0xc0009a3080) Reply frame received for 1\nI0425 00:12:51.107565 1088 log.go:172] (0xc0009a3080) (0xc0005a6000) Create stream\nI0425 00:12:51.107583 1088 log.go:172] (0xc0009a3080) (0xc0005a6000) Stream added, broadcasting: 3\nI0425 00:12:51.108686 1088 log.go:172] (0xc0009a3080) Reply frame received for 3\nI0425 00:12:51.108744 1088 log.go:172] (0xc0009a3080) (0xc000673680) Create stream\nI0425 00:12:51.108759 1088 log.go:172] (0xc0009a3080) (0xc000673680) Stream added, broadcasting: 5\nI0425 00:12:51.109770 1088 log.go:172] (0xc0009a3080) Reply frame received for 5\nI0425 00:12:51.191688 1088 log.go:172] (0xc0009a3080) Data frame received for 5\nI0425 00:12:51.191715 1088 log.go:172] (0xc000673680) (5) Data frame handling\nI0425 00:12:51.191730 1088 log.go:172] (0xc000673680) (5) Data frame sent\n+ nslookup clusterip-service\nI0425 00:12:51.198185 1088 log.go:172] (0xc0009a3080) Data frame received for 3\nI0425 00:12:51.198208 1088 log.go:172] (0xc0005a6000) (3) Data frame handling\nI0425 00:12:51.198225 1088 log.go:172] (0xc0005a6000) (3) Data frame sent\nI0425 00:12:51.199093 1088 log.go:172] (0xc0009a3080) Data frame received for 3\nI0425 00:12:51.199119 1088 log.go:172] (0xc0005a6000) (3) Data frame handling\nI0425 00:12:51.199140 1088 log.go:172] (0xc0005a6000) (3) Data frame sent\nI0425 00:12:51.199505 1088 log.go:172] (0xc0009a3080) Data frame received for 3\nI0425 00:12:51.199524 1088 log.go:172] (0xc0005a6000) (3) Data frame handling\nI0425 00:12:51.199631 1088 log.go:172] (0xc0009a3080) Data frame received for 5\nI0425 00:12:51.199645 1088 log.go:172] (0xc000673680) (5) Data frame handling\nI0425 00:12:51.201221 1088 log.go:172] (0xc0009a3080) Data frame received for 1\nI0425 00:12:51.201240 1088 log.go:172] (0xc0005a65a0) (1) Data frame handling\nI0425 00:12:51.201256 1088 log.go:172] (0xc0005a65a0) (1) Data frame sent\nI0425 00:12:51.201268 1088 log.go:172] (0xc0009a3080) (0xc0005a65a0) Stream removed, broadcasting: 1\nI0425 00:12:51.201295 1088 log.go:172] (0xc0009a3080) Go away received\nI0425 00:12:51.201560 1088 log.go:172] (0xc0009a3080) (0xc0005a65a0) Stream removed, broadcasting: 1\nI0425 00:12:51.201575 1088 log.go:172] (0xc0009a3080) (0xc0005a6000) Stream removed, broadcasting: 3\nI0425 00:12:51.201582 1088 log.go:172] (0xc0009a3080) (0xc000673680) Stream removed, broadcasting: 5\n" Apr 25 00:12:51.206: INFO: stdout: "Server:\t\t10.96.0.10\nAddress:\t10.96.0.10#53\n\nclusterip-service.services-4176.svc.cluster.local\tcanonical name = externalsvc.services-4176.svc.cluster.local.\nName:\texternalsvc.services-4176.svc.cluster.local\nAddress: 10.96.148.225\n\n" STEP: deleting ReplicationController externalsvc in namespace services-4176, will wait for the garbage collector to delete the pods Apr 25 00:12:51.264: INFO: Deleting ReplicationController externalsvc took: 5.715223ms Apr 25 00:12:51.365: INFO: Terminating ReplicationController externalsvc pods took: 100.336955ms Apr 25 00:13:03.094: INFO: Cleaning up the ClusterIP to ExternalName test service [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:13:03.105: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-4176" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 • [SLOW TEST:25.281 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from ClusterIP to ExternalName [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","total":275,"completed":116,"skipped":2073,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:13:03.128: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating projection with secret that has name projected-secret-test-map-8a8fffe8-11d2-4714-9aa7-0014dae518ca STEP: Creating a pod to test consume secrets Apr 25 00:13:03.211: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950" in namespace "projected-5919" to be "Succeeded or Failed" Apr 25 00:13:03.216: INFO: Pod "pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950": Phase="Pending", Reason="", readiness=false. Elapsed: 5.595486ms Apr 25 00:13:05.221: INFO: Pod "pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009787532s Apr 25 00:13:07.225: INFO: Pod "pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014265719s STEP: Saw pod success Apr 25 00:13:07.225: INFO: Pod "pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950" satisfied condition "Succeeded or Failed" Apr 25 00:13:07.228: INFO: Trying to get logs from node latest-worker pod pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950 container projected-secret-volume-test: STEP: delete the pod Apr 25 00:13:07.246: INFO: Waiting for pod pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950 to disappear Apr 25 00:13:07.250: INFO: Pod pod-projected-secrets-abf80de6-7b63-4953-b49a-f465f3827950 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:13:07.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5919" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":117,"skipped":2101,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:13:07.259: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a configMap. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ConfigMap STEP: Ensuring resource quota status captures configMap creation STEP: Deleting a ConfigMap STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:13:23.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-2681" for this suite. • [SLOW TEST:16.098 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a configMap. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","total":275,"completed":118,"skipped":2141,"failed":0} SS ------------------------------ [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:13:23.357: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename aggregator STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:76 Apr 25 00:13:23.494: INFO: >>> kubeConfig: /root/.kube/config [It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering the sample API server. Apr 25 00:13:24.768: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set Apr 25 00:13:26.950: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370404, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370404, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370404, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370404, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-76974b4fff\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:13:29.482: INFO: Waited 520.810723ms for the sample-apiserver to be ready to handle requests. [AfterEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:67 [AfterEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:13:29.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "aggregator-9432" for this suite. • [SLOW TEST:6.747 seconds] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]","total":275,"completed":119,"skipped":2143,"failed":0} [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:13:30.105: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating projection with secret that has name projected-secret-test-0e210ab4-d0ad-42aa-ac8f-9b3b2564abdf STEP: Creating a pod to test consume secrets Apr 25 00:13:30.188: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd" in namespace "projected-3188" to be "Succeeded or Failed" Apr 25 00:13:30.203: INFO: Pod "pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd": Phase="Pending", Reason="", readiness=false. Elapsed: 15.452766ms Apr 25 00:13:32.207: INFO: Pod "pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01917974s Apr 25 00:13:34.210: INFO: Pod "pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022855481s STEP: Saw pod success Apr 25 00:13:34.211: INFO: Pod "pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd" satisfied condition "Succeeded or Failed" Apr 25 00:13:34.213: INFO: Trying to get logs from node latest-worker2 pod pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd container projected-secret-volume-test: STEP: delete the pod Apr 25 00:13:34.305: INFO: Waiting for pod pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd to disappear Apr 25 00:13:34.311: INFO: Pod pod-projected-secrets-ae3213db-591a-4d50-837e-13d0098d7abd no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:13:34.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3188" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","total":275,"completed":120,"skipped":2143,"failed":0} ------------------------------ [sig-network] Services should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:13:34.318: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating service endpoint-test2 in namespace services-1384 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-1384 to expose endpoints map[] Apr 25 00:13:34.395: INFO: Get endpoints failed (5.112195ms elapsed, ignoring for 5s): endpoints "endpoint-test2" not found Apr 25 00:13:35.401: INFO: successfully validated that service endpoint-test2 in namespace services-1384 exposes endpoints map[] (1.011637665s elapsed) STEP: Creating pod pod1 in namespace services-1384 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-1384 to expose endpoints map[pod1:[80]] Apr 25 00:13:38.462: INFO: successfully validated that service endpoint-test2 in namespace services-1384 exposes endpoints map[pod1:[80]] (3.054665867s elapsed) STEP: Creating pod pod2 in namespace services-1384 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-1384 to expose endpoints map[pod1:[80] pod2:[80]] Apr 25 00:13:42.880: INFO: Unexpected endpoints: found map[bb6c6aac-e6ee-4004-86d7-30e7e703b549:[80]], expected map[pod1:[80] pod2:[80]] (4.41307299s elapsed, will retry) Apr 25 00:13:43.890: INFO: successfully validated that service endpoint-test2 in namespace services-1384 exposes endpoints map[pod1:[80] pod2:[80]] (5.423646404s elapsed) STEP: Deleting pod pod1 in namespace services-1384 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-1384 to expose endpoints map[pod2:[80]] Apr 25 00:13:45.294: INFO: successfully validated that service endpoint-test2 in namespace services-1384 exposes endpoints map[pod2:[80]] (1.395482437s elapsed) STEP: Deleting pod pod2 in namespace services-1384 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-1384 to expose endpoints map[] Apr 25 00:13:47.261: INFO: successfully validated that service endpoint-test2 in namespace services-1384 exposes endpoints map[] (1.962441778s elapsed) [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:13:47.476: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-1384" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 • [SLOW TEST:13.167 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods [Conformance]","total":275,"completed":121,"skipped":2143,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:13:47.485: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward api env vars Apr 25 00:13:47.561: INFO: Waiting up to 5m0s for pod "downward-api-3339a184-7eef-4aac-90e8-ebec015fa797" in namespace "downward-api-1447" to be "Succeeded or Failed" Apr 25 00:13:47.578: INFO: Pod "downward-api-3339a184-7eef-4aac-90e8-ebec015fa797": Phase="Pending", Reason="", readiness=false. Elapsed: 17.220382ms Apr 25 00:13:49.583: INFO: Pod "downward-api-3339a184-7eef-4aac-90e8-ebec015fa797": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021601843s Apr 25 00:13:51.587: INFO: Pod "downward-api-3339a184-7eef-4aac-90e8-ebec015fa797": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025503368s STEP: Saw pod success Apr 25 00:13:51.587: INFO: Pod "downward-api-3339a184-7eef-4aac-90e8-ebec015fa797" satisfied condition "Succeeded or Failed" Apr 25 00:13:51.590: INFO: Trying to get logs from node latest-worker pod downward-api-3339a184-7eef-4aac-90e8-ebec015fa797 container dapi-container: STEP: delete the pod Apr 25 00:13:51.622: INFO: Waiting for pod downward-api-3339a184-7eef-4aac-90e8-ebec015fa797 to disappear Apr 25 00:13:51.635: INFO: Pod downward-api-3339a184-7eef-4aac-90e8-ebec015fa797 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:13:51.635: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1447" for this suite. •{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","total":275,"completed":122,"skipped":2164,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:13:51.643: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [BeforeEach] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1288 STEP: creating an pod Apr 25 00:13:51.700: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config run logs-generator --image=us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 --namespace=kubectl-3526 -- logs-generator --log-lines-total 100 --run-duration 20s' Apr 25 00:13:51.807: INFO: stderr: "" Apr 25 00:13:51.807: INFO: stdout: "pod/logs-generator created\n" [It] should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Waiting for log generator to start. Apr 25 00:13:51.807: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] Apr 25 00:13:51.807: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-3526" to be "running and ready, or succeeded" Apr 25 00:13:51.814: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 7.112771ms Apr 25 00:13:53.817: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010122847s Apr 25 00:13:55.821: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 4.014089766s Apr 25 00:13:55.821: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" Apr 25 00:13:55.821: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] STEP: checking for a matching strings Apr 25 00:13:55.821: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-3526' Apr 25 00:13:55.935: INFO: stderr: "" Apr 25 00:13:55.935: INFO: stdout: "I0425 00:13:54.230746 1 logs_generator.go:76] 0 POST /api/v1/namespaces/kube-system/pods/pzf 574\nI0425 00:13:54.430872 1 logs_generator.go:76] 1 POST /api/v1/namespaces/ns/pods/rt2 230\nI0425 00:13:54.630983 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/4ll 300\nI0425 00:13:54.830929 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/99mp 287\nI0425 00:13:55.030990 1 logs_generator.go:76] 4 PUT /api/v1/namespaces/ns/pods/zzj 568\nI0425 00:13:55.230939 1 logs_generator.go:76] 5 GET /api/v1/namespaces/default/pods/rqrd 506\nI0425 00:13:55.430927 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/gxff 535\nI0425 00:13:55.630938 1 logs_generator.go:76] 7 POST /api/v1/namespaces/ns/pods/c7z 500\nI0425 00:13:55.830996 1 logs_generator.go:76] 8 GET /api/v1/namespaces/default/pods/9hz 586\n" STEP: limiting log lines Apr 25 00:13:55.935: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-3526 --tail=1' Apr 25 00:13:56.044: INFO: stderr: "" Apr 25 00:13:56.044: INFO: stdout: "I0425 00:13:56.030897 1 logs_generator.go:76] 9 POST /api/v1/namespaces/ns/pods/bbm4 376\n" Apr 25 00:13:56.044: INFO: got output "I0425 00:13:56.030897 1 logs_generator.go:76] 9 POST /api/v1/namespaces/ns/pods/bbm4 376\n" STEP: limiting log bytes Apr 25 00:13:56.044: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-3526 --limit-bytes=1' Apr 25 00:13:56.157: INFO: stderr: "" Apr 25 00:13:56.157: INFO: stdout: "I" Apr 25 00:13:56.157: INFO: got output "I" STEP: exposing timestamps Apr 25 00:13:56.157: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-3526 --tail=1 --timestamps' Apr 25 00:13:56.257: INFO: stderr: "" Apr 25 00:13:56.257: INFO: stdout: "2020-04-25T00:13:56.231003967Z I0425 00:13:56.230871 1 logs_generator.go:76] 10 GET /api/v1/namespaces/default/pods/pqr8 419\n" Apr 25 00:13:56.257: INFO: got output "2020-04-25T00:13:56.231003967Z I0425 00:13:56.230871 1 logs_generator.go:76] 10 GET /api/v1/namespaces/default/pods/pqr8 419\n" STEP: restricting to a time range Apr 25 00:13:58.757: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-3526 --since=1s' Apr 25 00:13:58.874: INFO: stderr: "" Apr 25 00:13:58.874: INFO: stdout: "I0425 00:13:58.030893 1 logs_generator.go:76] 19 GET /api/v1/namespaces/default/pods/4fx 598\nI0425 00:13:58.230931 1 logs_generator.go:76] 20 PUT /api/v1/namespaces/default/pods/d9zc 317\nI0425 00:13:58.430907 1 logs_generator.go:76] 21 POST /api/v1/namespaces/kube-system/pods/vw2w 218\nI0425 00:13:58.630901 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/kube-system/pods/stx 259\nI0425 00:13:58.830984 1 logs_generator.go:76] 23 GET /api/v1/namespaces/kube-system/pods/4t4q 255\n" Apr 25 00:13:58.875: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-3526 --since=24h' Apr 25 00:13:58.974: INFO: stderr: "" Apr 25 00:13:58.974: INFO: stdout: "I0425 00:13:54.230746 1 logs_generator.go:76] 0 POST /api/v1/namespaces/kube-system/pods/pzf 574\nI0425 00:13:54.430872 1 logs_generator.go:76] 1 POST /api/v1/namespaces/ns/pods/rt2 230\nI0425 00:13:54.630983 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/4ll 300\nI0425 00:13:54.830929 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/99mp 287\nI0425 00:13:55.030990 1 logs_generator.go:76] 4 PUT /api/v1/namespaces/ns/pods/zzj 568\nI0425 00:13:55.230939 1 logs_generator.go:76] 5 GET /api/v1/namespaces/default/pods/rqrd 506\nI0425 00:13:55.430927 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/gxff 535\nI0425 00:13:55.630938 1 logs_generator.go:76] 7 POST /api/v1/namespaces/ns/pods/c7z 500\nI0425 00:13:55.830996 1 logs_generator.go:76] 8 GET /api/v1/namespaces/default/pods/9hz 586\nI0425 00:13:56.030897 1 logs_generator.go:76] 9 POST /api/v1/namespaces/ns/pods/bbm4 376\nI0425 00:13:56.230871 1 logs_generator.go:76] 10 GET /api/v1/namespaces/default/pods/pqr8 419\nI0425 00:13:56.430944 1 logs_generator.go:76] 11 PUT /api/v1/namespaces/kube-system/pods/cm4x 355\nI0425 00:13:56.630972 1 logs_generator.go:76] 12 POST /api/v1/namespaces/kube-system/pods/r9vd 491\nI0425 00:13:56.830936 1 logs_generator.go:76] 13 POST /api/v1/namespaces/kube-system/pods/mnk 545\nI0425 00:13:57.030923 1 logs_generator.go:76] 14 PUT /api/v1/namespaces/ns/pods/w48j 272\nI0425 00:13:57.230926 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/default/pods/r8jz 380\nI0425 00:13:57.430883 1 logs_generator.go:76] 16 POST /api/v1/namespaces/kube-system/pods/x97 405\nI0425 00:13:57.630948 1 logs_generator.go:76] 17 GET /api/v1/namespaces/kube-system/pods/7sdl 249\nI0425 00:13:57.830967 1 logs_generator.go:76] 18 PUT /api/v1/namespaces/default/pods/sh9 248\nI0425 00:13:58.030893 1 logs_generator.go:76] 19 GET /api/v1/namespaces/default/pods/4fx 598\nI0425 00:13:58.230931 1 logs_generator.go:76] 20 PUT /api/v1/namespaces/default/pods/d9zc 317\nI0425 00:13:58.430907 1 logs_generator.go:76] 21 POST /api/v1/namespaces/kube-system/pods/vw2w 218\nI0425 00:13:58.630901 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/kube-system/pods/stx 259\nI0425 00:13:58.830984 1 logs_generator.go:76] 23 GET /api/v1/namespaces/kube-system/pods/4t4q 255\n" [AfterEach] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1294 Apr 25 00:13:58.974: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete pod logs-generator --namespace=kubectl-3526' Apr 25 00:14:01.162: INFO: stderr: "" Apr 25 00:14:01.162: INFO: stdout: "pod \"logs-generator\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:14:01.162: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-3526" for this suite. • [SLOW TEST:9.526 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1284 should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance]","total":275,"completed":123,"skipped":2187,"failed":0} SSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:14:01.169: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Apr 25 00:14:09.314: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:09.344: INFO: Pod pod-with-poststart-exec-hook still exists Apr 25 00:14:11.345: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:11.350: INFO: Pod pod-with-poststart-exec-hook still exists Apr 25 00:14:13.345: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:13.351: INFO: Pod pod-with-poststart-exec-hook still exists Apr 25 00:14:15.345: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:15.350: INFO: Pod pod-with-poststart-exec-hook still exists Apr 25 00:14:17.345: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:17.350: INFO: Pod pod-with-poststart-exec-hook still exists Apr 25 00:14:19.345: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:19.349: INFO: Pod pod-with-poststart-exec-hook still exists Apr 25 00:14:21.345: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:21.349: INFO: Pod pod-with-poststart-exec-hook still exists Apr 25 00:14:23.345: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 25 00:14:23.349: INFO: Pod pod-with-poststart-exec-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:14:23.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-5148" for this suite. • [SLOW TEST:22.188 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","total":275,"completed":124,"skipped":2192,"failed":0} SSSSS ------------------------------ [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:14:23.358: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating the pod Apr 25 00:14:27.975: INFO: Successfully updated pod "labelsupdate5ad17479-bc23-45da-bdff-9c7c325859d4" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:14:30.034: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-757" for this suite. • [SLOW TEST:6.684 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":275,"completed":125,"skipped":2197,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:14:30.043: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Given a ReplicationController is created STEP: When the matched label of one of its pods change Apr 25 00:14:30.121: INFO: Pod name pod-release: Found 0 pods out of 1 Apr 25 00:14:35.125: INFO: Pod name pod-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:14:36.137: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-6544" for this suite. • [SLOW TEST:6.104 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","total":275,"completed":126,"skipped":2211,"failed":0} SSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:14:36.146: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:84 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:99 STEP: Creating service test in namespace statefulset-6511 [It] should have a working scale subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating statefulset ss in namespace statefulset-6511 Apr 25 00:14:36.240: INFO: Found 0 stateful pods, waiting for 1 Apr 25 00:14:46.244: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: getting scale subresource STEP: updating a scale subresource STEP: verifying the statefulset Spec.Replicas was modified [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:110 Apr 25 00:14:46.267: INFO: Deleting all statefulset in ns statefulset-6511 Apr 25 00:14:46.274: INFO: Scaling statefulset ss to 0 Apr 25 00:14:56.356: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:14:56.359: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:14:56.375: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-6511" for this suite. • [SLOW TEST:20.236 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should have a working scale subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","total":275,"completed":127,"skipped":2216,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:14:56.383: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of same group and version but different kinds [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation Apr 25 00:14:56.424: INFO: >>> kubeConfig: /root/.kube/config Apr 25 00:14:59.421: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:15:10.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-3897" for this suite. • [SLOW TEST:14.120 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of same group and version but different kinds [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","total":275,"completed":128,"skipped":2241,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:15:10.504: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153 [It] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod Apr 25 00:15:10.548: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:15:20.331: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-7299" for this suite. • [SLOW TEST:9.878 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","total":275,"completed":129,"skipped":2273,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:15:20.383: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:15:20.932: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:15:22.944: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370520, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370520, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370521, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370520, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:15:25.991: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should deny crd creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering the crd webhook via the AdmissionRegistration API STEP: Creating a custom resource definition that should be denied by the webhook Apr 25 00:15:26.014: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:15:26.025: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-4279" for this suite. STEP: Destroying namespace "webhook-4279-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:5.719 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should deny crd creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":275,"completed":130,"skipped":2296,"failed":0} SSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:15:26.102: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test override arguments Apr 25 00:15:26.156: INFO: Waiting up to 5m0s for pod "client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc" in namespace "containers-8434" to be "Succeeded or Failed" Apr 25 00:15:26.171: INFO: Pod "client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc": Phase="Pending", Reason="", readiness=false. Elapsed: 15.114995ms Apr 25 00:15:28.176: INFO: Pod "client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019687958s Apr 25 00:15:30.180: INFO: Pod "client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023850678s STEP: Saw pod success Apr 25 00:15:30.180: INFO: Pod "client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc" satisfied condition "Succeeded or Failed" Apr 25 00:15:30.183: INFO: Trying to get logs from node latest-worker pod client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc container test-container: STEP: delete the pod Apr 25 00:15:30.227: INFO: Waiting for pod client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc to disappear Apr 25 00:15:30.274: INFO: Pod client-containers-2dc49c86-4cec-4b16-819a-b6b70ef90bcc no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:15:30.274: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-8434" for this suite. •{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]","total":275,"completed":131,"skipped":2305,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:15:30.285: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the rc1 STEP: create the rc2 STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well STEP: delete the rc simpletest-rc-to-be-deleted STEP: wait for the rc to be deleted STEP: Gathering metrics W0425 00:15:41.753591 8 metrics_grabber.go:84] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 25 00:15:41.753: INFO: For apiserver_request_total: For apiserver_request_latency_seconds: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:15:41.753: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-5466" for this suite. • [SLOW TEST:11.475 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","total":275,"completed":132,"skipped":2342,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-network] Service endpoints latency should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:15:41.761: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svc-latency STEP: Waiting for a default service account to be provisioned in namespace [It] should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:15:41.892: INFO: >>> kubeConfig: /root/.kube/config STEP: creating replication controller svc-latency-rc in namespace svc-latency-9871 I0425 00:15:41.907267 8 runners.go:190] Created replication controller with name: svc-latency-rc, namespace: svc-latency-9871, replica count: 1 I0425 00:15:42.957803 8 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0425 00:15:43.957984 8 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0425 00:15:44.958183 8 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Apr 25 00:15:45.082: INFO: Created: latency-svc-xxz6g Apr 25 00:15:45.096: INFO: Got endpoints: latency-svc-xxz6g [37.658381ms] Apr 25 00:15:45.118: INFO: Created: latency-svc-f57pm Apr 25 00:15:45.130: INFO: Got endpoints: latency-svc-f57pm [34.103851ms] Apr 25 00:15:45.167: INFO: Created: latency-svc-9wv6n Apr 25 00:15:45.172: INFO: Got endpoints: latency-svc-9wv6n [75.241102ms] Apr 25 00:15:45.191: INFO: Created: latency-svc-r44qn Apr 25 00:15:45.208: INFO: Got endpoints: latency-svc-r44qn [111.240815ms] Apr 25 00:15:45.239: INFO: Created: latency-svc-hrcms Apr 25 00:15:45.256: INFO: Got endpoints: latency-svc-hrcms [158.720962ms] Apr 25 00:15:45.286: INFO: Created: latency-svc-24w96 Apr 25 00:15:45.304: INFO: Created: latency-svc-pvqpp Apr 25 00:15:45.304: INFO: Got endpoints: latency-svc-24w96 [207.249702ms] Apr 25 00:15:45.317: INFO: Got endpoints: latency-svc-pvqpp [220.424911ms] Apr 25 00:15:45.334: INFO: Created: latency-svc-s9kv6 Apr 25 00:15:45.347: INFO: Got endpoints: latency-svc-s9kv6 [250.058565ms] Apr 25 00:15:45.364: INFO: Created: latency-svc-7sdjm Apr 25 00:15:45.377: INFO: Got endpoints: latency-svc-7sdjm [280.024193ms] Apr 25 00:15:45.419: INFO: Created: latency-svc-jnt8g Apr 25 00:15:45.438: INFO: Got endpoints: latency-svc-jnt8g [341.333104ms] Apr 25 00:15:45.439: INFO: Created: latency-svc-s5lsg Apr 25 00:15:45.449: INFO: Got endpoints: latency-svc-s5lsg [351.843596ms] Apr 25 00:15:45.466: INFO: Created: latency-svc-gmvg8 Apr 25 00:15:45.485: INFO: Got endpoints: latency-svc-gmvg8 [388.139416ms] Apr 25 00:15:45.501: INFO: Created: latency-svc-l6rg4 Apr 25 00:15:45.515: INFO: Got endpoints: latency-svc-l6rg4 [417.860337ms] Apr 25 00:15:45.538: INFO: Created: latency-svc-n6zlw Apr 25 00:15:45.549: INFO: Got endpoints: latency-svc-n6zlw [452.010809ms] Apr 25 00:15:45.569: INFO: Created: latency-svc-wnk4c Apr 25 00:15:45.579: INFO: Got endpoints: latency-svc-wnk4c [481.892806ms] Apr 25 00:15:45.598: INFO: Created: latency-svc-fmbhr Apr 25 00:15:45.609: INFO: Got endpoints: latency-svc-fmbhr [511.771117ms] Apr 25 00:15:45.628: INFO: Created: latency-svc-mv9j9 Apr 25 00:15:45.663: INFO: Got endpoints: latency-svc-mv9j9 [533.07192ms] Apr 25 00:15:45.676: INFO: Created: latency-svc-9jjsb Apr 25 00:15:45.706: INFO: Got endpoints: latency-svc-9jjsb [534.807911ms] Apr 25 00:15:45.735: INFO: Created: latency-svc-hwcfr Apr 25 00:15:45.747: INFO: Got endpoints: latency-svc-hwcfr [538.78532ms] Apr 25 00:15:45.796: INFO: Created: latency-svc-vl668 Apr 25 00:15:45.814: INFO: Got endpoints: latency-svc-vl668 [558.10577ms] Apr 25 00:15:45.814: INFO: Created: latency-svc-ll6kx Apr 25 00:15:45.826: INFO: Got endpoints: latency-svc-ll6kx [522.197384ms] Apr 25 00:15:45.844: INFO: Created: latency-svc-rtqrw Apr 25 00:15:45.857: INFO: Got endpoints: latency-svc-rtqrw [539.852715ms] Apr 25 00:15:45.874: INFO: Created: latency-svc-cdn46 Apr 25 00:15:45.887: INFO: Got endpoints: latency-svc-cdn46 [539.436061ms] Apr 25 00:15:45.927: INFO: Created: latency-svc-wfxfw Apr 25 00:15:45.947: INFO: Created: latency-svc-rn5b9 Apr 25 00:15:45.947: INFO: Got endpoints: latency-svc-wfxfw [570.023691ms] Apr 25 00:15:45.958: INFO: Got endpoints: latency-svc-rn5b9 [519.499316ms] Apr 25 00:15:45.994: INFO: Created: latency-svc-6nw79 Apr 25 00:15:46.024: INFO: Got endpoints: latency-svc-6nw79 [575.317977ms] Apr 25 00:15:46.065: INFO: Created: latency-svc-hkh66 Apr 25 00:15:46.078: INFO: Got endpoints: latency-svc-hkh66 [593.219519ms] Apr 25 00:15:46.102: INFO: Created: latency-svc-9jgg6 Apr 25 00:15:46.115: INFO: Got endpoints: latency-svc-9jgg6 [599.686657ms] Apr 25 00:15:46.132: INFO: Created: latency-svc-69z2b Apr 25 00:15:46.149: INFO: Got endpoints: latency-svc-69z2b [599.874146ms] Apr 25 00:15:46.203: INFO: Created: latency-svc-2b7mq Apr 25 00:15:46.221: INFO: Got endpoints: latency-svc-2b7mq [642.343786ms] Apr 25 00:15:46.221: INFO: Created: latency-svc-jbt6p Apr 25 00:15:46.233: INFO: Got endpoints: latency-svc-jbt6p [623.567267ms] Apr 25 00:15:46.245: INFO: Created: latency-svc-hbb5n Apr 25 00:15:46.256: INFO: Got endpoints: latency-svc-hbb5n [593.068503ms] Apr 25 00:15:46.283: INFO: Created: latency-svc-vvmrk Apr 25 00:15:46.382: INFO: Got endpoints: latency-svc-vvmrk [675.917397ms] Apr 25 00:15:46.402: INFO: Created: latency-svc-4tjr7 Apr 25 00:15:46.418: INFO: Got endpoints: latency-svc-4tjr7 [671.302615ms] Apr 25 00:15:46.450: INFO: Created: latency-svc-8fw4z Apr 25 00:15:46.466: INFO: Got endpoints: latency-svc-8fw4z [652.542445ms] Apr 25 00:15:46.514: INFO: Created: latency-svc-nhxd5 Apr 25 00:15:46.528: INFO: Got endpoints: latency-svc-nhxd5 [701.10252ms] Apr 25 00:15:46.552: INFO: Created: latency-svc-qfthr Apr 25 00:15:46.570: INFO: Got endpoints: latency-svc-qfthr [712.292117ms] Apr 25 00:15:46.590: INFO: Created: latency-svc-9d7n7 Apr 25 00:15:46.599: INFO: Got endpoints: latency-svc-9d7n7 [712.736553ms] Apr 25 00:15:46.646: INFO: Created: latency-svc-gdhn9 Apr 25 00:15:46.653: INFO: Got endpoints: latency-svc-gdhn9 [706.254043ms] Apr 25 00:15:46.671: INFO: Created: latency-svc-dkzxk Apr 25 00:15:46.683: INFO: Got endpoints: latency-svc-dkzxk [724.631865ms] Apr 25 00:15:46.993: INFO: Created: latency-svc-86kgr Apr 25 00:15:47.002: INFO: Got endpoints: latency-svc-86kgr [977.470324ms] Apr 25 00:15:47.445: INFO: Created: latency-svc-x64gp Apr 25 00:15:47.462: INFO: Got endpoints: latency-svc-x64gp [1.383804044s] Apr 25 00:15:47.491: INFO: Created: latency-svc-fm68w Apr 25 00:15:47.516: INFO: Got endpoints: latency-svc-fm68w [1.401126399s] Apr 25 00:15:47.574: INFO: Created: latency-svc-cbk72 Apr 25 00:15:47.600: INFO: Created: latency-svc-mlljn Apr 25 00:15:47.600: INFO: Got endpoints: latency-svc-cbk72 [1.451307929s] Apr 25 00:15:47.616: INFO: Got endpoints: latency-svc-mlljn [1.394515011s] Apr 25 00:15:48.179: INFO: Created: latency-svc-8f275 Apr 25 00:15:48.212: INFO: Got endpoints: latency-svc-8f275 [1.979246739s] Apr 25 00:15:49.027: INFO: Created: latency-svc-qh885 Apr 25 00:15:49.059: INFO: Got endpoints: latency-svc-qh885 [2.803069522s] Apr 25 00:15:49.182: INFO: Created: latency-svc-c5rjr Apr 25 00:15:49.248: INFO: Got endpoints: latency-svc-c5rjr [2.865567822s] Apr 25 00:15:49.394: INFO: Created: latency-svc-pmgxm Apr 25 00:15:49.432: INFO: Got endpoints: latency-svc-pmgxm [3.013855792s] Apr 25 00:15:49.484: INFO: Created: latency-svc-qxnc2 Apr 25 00:15:49.513: INFO: Got endpoints: latency-svc-qxnc2 [3.046578069s] Apr 25 00:15:49.513: INFO: Created: latency-svc-b8hlk Apr 25 00:15:49.554: INFO: Got endpoints: latency-svc-b8hlk [3.026514319s] Apr 25 00:15:49.622: INFO: Created: latency-svc-g99r8 Apr 25 00:15:49.658: INFO: Created: latency-svc-962wm Apr 25 00:15:49.658: INFO: Got endpoints: latency-svc-g99r8 [3.088330257s] Apr 25 00:15:49.684: INFO: Got endpoints: latency-svc-962wm [3.084202767s] Apr 25 00:15:49.716: INFO: Created: latency-svc-bxvsl Apr 25 00:15:49.775: INFO: Got endpoints: latency-svc-bxvsl [3.12164158s] Apr 25 00:15:49.825: INFO: Created: latency-svc-w2kpr Apr 25 00:15:49.839: INFO: Got endpoints: latency-svc-w2kpr [3.156567769s] Apr 25 00:15:49.903: INFO: Created: latency-svc-zjhg8 Apr 25 00:15:49.915: INFO: Got endpoints: latency-svc-zjhg8 [2.913593858s] Apr 25 00:15:49.951: INFO: Created: latency-svc-z95m5 Apr 25 00:15:49.964: INFO: Got endpoints: latency-svc-z95m5 [2.501135115s] Apr 25 00:15:49.982: INFO: Created: latency-svc-x9pbv Apr 25 00:15:49.994: INFO: Got endpoints: latency-svc-x9pbv [2.477581247s] Apr 25 00:15:50.053: INFO: Created: latency-svc-5v9d6 Apr 25 00:15:50.059: INFO: Got endpoints: latency-svc-5v9d6 [2.458945531s] Apr 25 00:15:50.131: INFO: Created: latency-svc-857pc Apr 25 00:15:50.179: INFO: Got endpoints: latency-svc-857pc [2.563043666s] Apr 25 00:15:50.240: INFO: Created: latency-svc-rw7jj Apr 25 00:15:50.252: INFO: Got endpoints: latency-svc-rw7jj [2.039748673s] Apr 25 00:15:50.269: INFO: Created: latency-svc-d4zxc Apr 25 00:15:50.292: INFO: Got endpoints: latency-svc-d4zxc [1.232552198s] Apr 25 00:15:50.300: INFO: Created: latency-svc-nf2nn Apr 25 00:15:50.312: INFO: Got endpoints: latency-svc-nf2nn [1.064205121s] Apr 25 00:15:50.329: INFO: Created: latency-svc-glsvt Apr 25 00:15:50.343: INFO: Got endpoints: latency-svc-glsvt [910.341677ms] Apr 25 00:15:50.359: INFO: Created: latency-svc-wrzkw Apr 25 00:15:50.373: INFO: Got endpoints: latency-svc-wrzkw [859.502334ms] Apr 25 00:15:50.424: INFO: Created: latency-svc-4h4ch Apr 25 00:15:50.436: INFO: Got endpoints: latency-svc-4h4ch [881.995048ms] Apr 25 00:15:50.437: INFO: Created: latency-svc-75c2m Apr 25 00:15:50.461: INFO: Got endpoints: latency-svc-75c2m [803.006497ms] Apr 25 00:15:50.485: INFO: Created: latency-svc-mzrc9 Apr 25 00:15:50.499: INFO: Got endpoints: latency-svc-mzrc9 [814.873231ms] Apr 25 00:15:50.521: INFO: Created: latency-svc-sdh2v Apr 25 00:15:50.574: INFO: Got endpoints: latency-svc-sdh2v [799.019339ms] Apr 25 00:15:50.576: INFO: Created: latency-svc-sjtkr Apr 25 00:15:50.581: INFO: Got endpoints: latency-svc-sjtkr [741.4314ms] Apr 25 00:15:50.600: INFO: Created: latency-svc-nflwj Apr 25 00:15:50.617: INFO: Got endpoints: latency-svc-nflwj [701.33634ms] Apr 25 00:15:50.635: INFO: Created: latency-svc-glsl5 Apr 25 00:15:50.647: INFO: Got endpoints: latency-svc-glsl5 [683.007244ms] Apr 25 00:15:50.671: INFO: Created: latency-svc-xqkw5 Apr 25 00:15:50.693: INFO: Got endpoints: latency-svc-xqkw5 [699.865185ms] Apr 25 00:15:50.725: INFO: Created: latency-svc-qn9ms Apr 25 00:15:50.736: INFO: Got endpoints: latency-svc-qn9ms [676.894778ms] Apr 25 00:15:50.761: INFO: Created: latency-svc-xmf69 Apr 25 00:15:50.779: INFO: Got endpoints: latency-svc-xmf69 [599.955655ms] Apr 25 00:15:50.851: INFO: Created: latency-svc-sm49s Apr 25 00:15:50.856: INFO: Got endpoints: latency-svc-sm49s [604.46694ms] Apr 25 00:15:50.941: INFO: Created: latency-svc-98tz9 Apr 25 00:15:51.059: INFO: Got endpoints: latency-svc-98tz9 [766.342863ms] Apr 25 00:15:51.115: INFO: Created: latency-svc-m8gw7 Apr 25 00:15:51.146: INFO: Got endpoints: latency-svc-m8gw7 [833.354797ms] Apr 25 00:15:51.199: INFO: Created: latency-svc-p9q8g Apr 25 00:15:51.364: INFO: Got endpoints: latency-svc-p9q8g [1.021576481s] Apr 25 00:15:51.373: INFO: Created: latency-svc-5g2k2 Apr 25 00:15:51.392: INFO: Created: latency-svc-bg4j6 Apr 25 00:15:51.392: INFO: Got endpoints: latency-svc-5g2k2 [1.018867753s] Apr 25 00:15:51.397: INFO: Got endpoints: latency-svc-bg4j6 [960.35144ms] Apr 25 00:15:51.416: INFO: Created: latency-svc-s99dp Apr 25 00:15:51.427: INFO: Got endpoints: latency-svc-s99dp [965.663068ms] Apr 25 00:15:51.541: INFO: Created: latency-svc-9fvkl Apr 25 00:15:51.560: INFO: Created: latency-svc-djzhd Apr 25 00:15:51.560: INFO: Got endpoints: latency-svc-9fvkl [1.061218672s] Apr 25 00:15:51.575: INFO: Got endpoints: latency-svc-djzhd [1.001211529s] Apr 25 00:15:51.602: INFO: Created: latency-svc-75vhs Apr 25 00:15:51.646: INFO: Got endpoints: latency-svc-75vhs [1.065065173s] Apr 25 00:15:51.666: INFO: Created: latency-svc-xpw5q Apr 25 00:15:51.683: INFO: Got endpoints: latency-svc-xpw5q [1.066109125s] Apr 25 00:15:51.697: INFO: Created: latency-svc-88d4g Apr 25 00:15:51.707: INFO: Got endpoints: latency-svc-88d4g [1.0607628s] Apr 25 00:15:51.720: INFO: Created: latency-svc-f8zrp Apr 25 00:15:51.745: INFO: Got endpoints: latency-svc-f8zrp [1.051503868s] Apr 25 00:15:51.795: INFO: Created: latency-svc-5vs9n Apr 25 00:15:51.804: INFO: Got endpoints: latency-svc-5vs9n [1.068035649s] Apr 25 00:15:51.829: INFO: Created: latency-svc-q6dxj Apr 25 00:15:51.845: INFO: Got endpoints: latency-svc-q6dxj [1.066295014s] Apr 25 00:15:51.870: INFO: Created: latency-svc-bktrn Apr 25 00:15:51.927: INFO: Got endpoints: latency-svc-bktrn [1.070682033s] Apr 25 00:15:51.937: INFO: Created: latency-svc-lzjk7 Apr 25 00:15:51.954: INFO: Got endpoints: latency-svc-lzjk7 [895.444923ms] Apr 25 00:15:51.973: INFO: Created: latency-svc-mlbf7 Apr 25 00:15:51.990: INFO: Got endpoints: latency-svc-mlbf7 [844.387022ms] Apr 25 00:15:52.010: INFO: Created: latency-svc-4nb6m Apr 25 00:15:52.026: INFO: Got endpoints: latency-svc-4nb6m [661.938139ms] Apr 25 00:15:52.111: INFO: Created: latency-svc-zjhtb Apr 25 00:15:52.128: INFO: Got endpoints: latency-svc-zjhtb [736.747644ms] Apr 25 00:15:52.152: INFO: Created: latency-svc-qbtfx Apr 25 00:15:52.170: INFO: Got endpoints: latency-svc-qbtfx [773.114404ms] Apr 25 00:15:52.202: INFO: Created: latency-svc-j4bdn Apr 25 00:15:52.218: INFO: Got endpoints: latency-svc-j4bdn [790.917752ms] Apr 25 00:15:52.237: INFO: Created: latency-svc-4dnb4 Apr 25 00:15:52.252: INFO: Got endpoints: latency-svc-4dnb4 [692.63575ms] Apr 25 00:15:52.273: INFO: Created: latency-svc-gfgkn Apr 25 00:15:52.288: INFO: Got endpoints: latency-svc-gfgkn [712.946845ms] Apr 25 00:15:52.328: INFO: Created: latency-svc-564j7 Apr 25 00:15:52.350: INFO: Got endpoints: latency-svc-564j7 [704.254844ms] Apr 25 00:15:52.351: INFO: Created: latency-svc-97nn6 Apr 25 00:15:52.374: INFO: Got endpoints: latency-svc-97nn6 [691.410826ms] Apr 25 00:15:52.399: INFO: Created: latency-svc-mmvzp Apr 25 00:15:52.408: INFO: Got endpoints: latency-svc-mmvzp [700.44658ms] Apr 25 00:15:52.423: INFO: Created: latency-svc-6bpr2 Apr 25 00:15:52.455: INFO: Got endpoints: latency-svc-6bpr2 [710.010344ms] Apr 25 00:15:52.477: INFO: Created: latency-svc-dvxpq Apr 25 00:15:52.507: INFO: Got endpoints: latency-svc-dvxpq [702.155633ms] Apr 25 00:15:52.531: INFO: Created: latency-svc-2xldx Apr 25 00:15:52.542: INFO: Got endpoints: latency-svc-2xldx [696.090214ms] Apr 25 00:15:52.580: INFO: Created: latency-svc-58w9l Apr 25 00:15:52.583: INFO: Got endpoints: latency-svc-58w9l [656.105745ms] Apr 25 00:15:52.603: INFO: Created: latency-svc-gsbv9 Apr 25 00:15:52.633: INFO: Got endpoints: latency-svc-gsbv9 [678.73167ms] Apr 25 00:15:52.663: INFO: Created: latency-svc-6kfln Apr 25 00:15:52.679: INFO: Got endpoints: latency-svc-6kfln [688.955274ms] Apr 25 00:15:52.711: INFO: Created: latency-svc-gqgfw Apr 25 00:15:52.721: INFO: Got endpoints: latency-svc-gqgfw [694.448542ms] Apr 25 00:15:52.735: INFO: Created: latency-svc-rjf6s Apr 25 00:15:52.745: INFO: Got endpoints: latency-svc-rjf6s [616.366769ms] Apr 25 00:15:52.764: INFO: Created: latency-svc-wfn5j Apr 25 00:15:52.780: INFO: Got endpoints: latency-svc-wfn5j [609.73047ms] Apr 25 00:15:52.819: INFO: Created: latency-svc-zn2tx Apr 25 00:15:52.843: INFO: Created: latency-svc-ctx6m Apr 25 00:15:52.844: INFO: Got endpoints: latency-svc-zn2tx [626.376531ms] Apr 25 00:15:52.852: INFO: Got endpoints: latency-svc-ctx6m [599.11693ms] Apr 25 00:15:52.866: INFO: Created: latency-svc-cbllc Apr 25 00:15:52.891: INFO: Got endpoints: latency-svc-cbllc [603.056198ms] Apr 25 00:15:52.946: INFO: Created: latency-svc-sxvsm Apr 25 00:15:52.954: INFO: Got endpoints: latency-svc-sxvsm [603.391998ms] Apr 25 00:15:52.981: INFO: Created: latency-svc-sck2s Apr 25 00:15:52.995: INFO: Got endpoints: latency-svc-sck2s [620.656159ms] Apr 25 00:15:53.035: INFO: Created: latency-svc-xdjp8 Apr 25 00:15:53.095: INFO: Got endpoints: latency-svc-xdjp8 [686.863342ms] Apr 25 00:15:53.113: INFO: Created: latency-svc-b57pz Apr 25 00:15:53.128: INFO: Got endpoints: latency-svc-b57pz [673.087368ms] Apr 25 00:15:53.161: INFO: Created: latency-svc-pndr7 Apr 25 00:15:53.176: INFO: Got endpoints: latency-svc-pndr7 [669.462041ms] Apr 25 00:15:53.191: INFO: Created: latency-svc-vtv7q Apr 25 00:15:53.215: INFO: Got endpoints: latency-svc-vtv7q [673.082458ms] Apr 25 00:15:53.292: INFO: Created: latency-svc-hpxck Apr 25 00:15:53.308: INFO: Got endpoints: latency-svc-hpxck [724.824829ms] Apr 25 00:15:53.346: INFO: Created: latency-svc-dql5m Apr 25 00:15:53.356: INFO: Got endpoints: latency-svc-dql5m [723.128006ms] Apr 25 00:15:53.376: INFO: Created: latency-svc-9pmd7 Apr 25 00:15:53.386: INFO: Got endpoints: latency-svc-9pmd7 [706.664204ms] Apr 25 00:15:53.401: INFO: Created: latency-svc-2r6fr Apr 25 00:15:53.410: INFO: Got endpoints: latency-svc-2r6fr [688.728469ms] Apr 25 00:15:53.484: INFO: Created: latency-svc-qc9ww Apr 25 00:15:53.487: INFO: Got endpoints: latency-svc-qc9ww [742.389421ms] Apr 25 00:15:53.508: INFO: Created: latency-svc-njx4d Apr 25 00:15:53.523: INFO: Got endpoints: latency-svc-njx4d [743.255332ms] Apr 25 00:15:53.545: INFO: Created: latency-svc-92fng Apr 25 00:15:53.558: INFO: Got endpoints: latency-svc-92fng [714.084722ms] Apr 25 00:15:53.575: INFO: Created: latency-svc-fdthg Apr 25 00:15:53.675: INFO: Got endpoints: latency-svc-fdthg [823.673829ms] Apr 25 00:15:54.042: INFO: Created: latency-svc-hm594 Apr 25 00:15:54.080: INFO: Created: latency-svc-nf6vz Apr 25 00:15:54.081: INFO: Got endpoints: latency-svc-hm594 [1.18965897s] Apr 25 00:15:54.116: INFO: Got endpoints: latency-svc-nf6vz [1.162142395s] Apr 25 00:15:54.293: INFO: Created: latency-svc-kn2kb Apr 25 00:15:54.307: INFO: Got endpoints: latency-svc-kn2kb [1.312109336s] Apr 25 00:15:54.464: INFO: Created: latency-svc-tpfv6 Apr 25 00:15:54.482: INFO: Got endpoints: latency-svc-tpfv6 [1.387492572s] Apr 25 00:15:54.640: INFO: Created: latency-svc-db6s8 Apr 25 00:15:54.650: INFO: Got endpoints: latency-svc-db6s8 [1.522004579s] Apr 25 00:15:54.686: INFO: Created: latency-svc-4r2gl Apr 25 00:15:54.710: INFO: Got endpoints: latency-svc-4r2gl [1.533542009s] Apr 25 00:15:54.857: INFO: Created: latency-svc-4zldk Apr 25 00:15:54.872: INFO: Got endpoints: latency-svc-4zldk [1.657097152s] Apr 25 00:15:54.903: INFO: Created: latency-svc-gqnwk Apr 25 00:15:54.926: INFO: Got endpoints: latency-svc-gqnwk [1.618464138s] Apr 25 00:15:55.101: INFO: Created: latency-svc-wf99g Apr 25 00:15:55.149: INFO: Got endpoints: latency-svc-wf99g [1.793166483s] Apr 25 00:15:55.265: INFO: Created: latency-svc-bsg57 Apr 25 00:15:55.284: INFO: Got endpoints: latency-svc-bsg57 [1.89761547s] Apr 25 00:15:55.330: INFO: Created: latency-svc-6dbq8 Apr 25 00:15:55.461: INFO: Got endpoints: latency-svc-6dbq8 [2.051635026s] Apr 25 00:15:55.557: INFO: Created: latency-svc-qczxp Apr 25 00:15:55.711: INFO: Got endpoints: latency-svc-qczxp [2.224104834s] Apr 25 00:15:55.713: INFO: Created: latency-svc-z2nw4 Apr 25 00:15:55.721: INFO: Got endpoints: latency-svc-z2nw4 [2.197558643s] Apr 25 00:15:55.737: INFO: Created: latency-svc-6pz5h Apr 25 00:15:55.751: INFO: Got endpoints: latency-svc-6pz5h [2.192488823s] Apr 25 00:15:55.768: INFO: Created: latency-svc-7vpqx Apr 25 00:15:55.781: INFO: Got endpoints: latency-svc-7vpqx [2.105354656s] Apr 25 00:15:55.851: INFO: Created: latency-svc-qgv6g Apr 25 00:15:55.865: INFO: Got endpoints: latency-svc-qgv6g [1.783441456s] Apr 25 00:15:55.894: INFO: Created: latency-svc-4n49r Apr 25 00:15:55.908: INFO: Got endpoints: latency-svc-4n49r [1.791987519s] Apr 25 00:15:55.936: INFO: Created: latency-svc-8pqbd Apr 25 00:15:55.963: INFO: Got endpoints: latency-svc-8pqbd [1.655753087s] Apr 25 00:15:55.978: INFO: Created: latency-svc-5b7ms Apr 25 00:15:56.002: INFO: Got endpoints: latency-svc-5b7ms [1.519741038s] Apr 25 00:15:56.044: INFO: Created: latency-svc-df58p Apr 25 00:15:56.058: INFO: Got endpoints: latency-svc-df58p [1.407259202s] Apr 25 00:15:56.107: INFO: Created: latency-svc-54hhx Apr 25 00:15:56.134: INFO: Created: latency-svc-nslqx Apr 25 00:15:56.134: INFO: Got endpoints: latency-svc-54hhx [1.424452469s] Apr 25 00:15:56.148: INFO: Got endpoints: latency-svc-nslqx [1.275668992s] Apr 25 00:15:56.188: INFO: Created: latency-svc-57n24 Apr 25 00:15:56.202: INFO: Got endpoints: latency-svc-57n24 [1.275257928s] Apr 25 00:15:56.239: INFO: Created: latency-svc-2ldgn Apr 25 00:15:56.261: INFO: Got endpoints: latency-svc-2ldgn [1.111250216s] Apr 25 00:15:56.261: INFO: Created: latency-svc-z2lb9 Apr 25 00:15:56.272: INFO: Got endpoints: latency-svc-z2lb9 [988.766673ms] Apr 25 00:15:56.290: INFO: Created: latency-svc-497xr Apr 25 00:15:56.302: INFO: Got endpoints: latency-svc-497xr [840.560074ms] Apr 25 00:15:56.319: INFO: Created: latency-svc-8vrqm Apr 25 00:15:56.332: INFO: Got endpoints: latency-svc-8vrqm [620.82082ms] Apr 25 00:15:56.370: INFO: Created: latency-svc-6h5mx Apr 25 00:15:57.060: INFO: Created: latency-svc-8j94k Apr 25 00:15:57.068: INFO: Got endpoints: latency-svc-6h5mx [1.347485404s] Apr 25 00:15:57.070: INFO: Got endpoints: latency-svc-8j94k [1.318987784s] Apr 25 00:15:57.099: INFO: Created: latency-svc-pq8fd Apr 25 00:15:57.117: INFO: Got endpoints: latency-svc-pq8fd [1.336505429s] Apr 25 00:15:57.141: INFO: Created: latency-svc-mxsft Apr 25 00:15:57.155: INFO: Got endpoints: latency-svc-mxsft [1.290136371s] Apr 25 00:15:57.231: INFO: Created: latency-svc-rjxl8 Apr 25 00:15:57.243: INFO: Got endpoints: latency-svc-rjxl8 [1.335424011s] Apr 25 00:15:57.267: INFO: Created: latency-svc-l7kps Apr 25 00:15:57.279: INFO: Got endpoints: latency-svc-l7kps [1.316074483s] Apr 25 00:15:57.309: INFO: Created: latency-svc-rm4hc Apr 25 00:15:57.352: INFO: Got endpoints: latency-svc-rm4hc [1.349860707s] Apr 25 00:15:57.387: INFO: Created: latency-svc-gnfjg Apr 25 00:15:57.399: INFO: Got endpoints: latency-svc-gnfjg [1.341729543s] Apr 25 00:15:57.423: INFO: Created: latency-svc-j6w2m Apr 25 00:15:57.436: INFO: Got endpoints: latency-svc-j6w2m [1.301340133s] Apr 25 00:15:57.484: INFO: Created: latency-svc-gbxjl Apr 25 00:15:57.507: INFO: Created: latency-svc-hpr74 Apr 25 00:15:57.507: INFO: Got endpoints: latency-svc-gbxjl [1.359087256s] Apr 25 00:15:57.524: INFO: Got endpoints: latency-svc-hpr74 [1.321937649s] Apr 25 00:15:57.548: INFO: Created: latency-svc-l9kgh Apr 25 00:15:57.560: INFO: Got endpoints: latency-svc-l9kgh [1.298657684s] Apr 25 00:15:57.572: INFO: Created: latency-svc-l8lpl Apr 25 00:15:57.640: INFO: Got endpoints: latency-svc-l8lpl [1.367225857s] Apr 25 00:15:57.642: INFO: Created: latency-svc-jndqf Apr 25 00:15:57.649: INFO: Got endpoints: latency-svc-jndqf [1.347542358s] Apr 25 00:15:57.675: INFO: Created: latency-svc-x4crk Apr 25 00:15:57.694: INFO: Got endpoints: latency-svc-x4crk [1.361655877s] Apr 25 00:15:57.711: INFO: Created: latency-svc-65f2l Apr 25 00:15:57.727: INFO: Got endpoints: latency-svc-65f2l [659.097494ms] Apr 25 00:15:57.808: INFO: Created: latency-svc-g85cb Apr 25 00:15:57.811: INFO: Got endpoints: latency-svc-g85cb [741.578538ms] Apr 25 00:15:57.836: INFO: Created: latency-svc-djm9l Apr 25 00:15:57.849: INFO: Got endpoints: latency-svc-djm9l [731.672296ms] Apr 25 00:15:57.867: INFO: Created: latency-svc-kpmvx Apr 25 00:15:57.879: INFO: Got endpoints: latency-svc-kpmvx [723.622401ms] Apr 25 00:15:57.903: INFO: Created: latency-svc-xngkq Apr 25 00:15:57.945: INFO: Got endpoints: latency-svc-xngkq [701.873849ms] Apr 25 00:15:57.947: INFO: Created: latency-svc-292br Apr 25 00:15:57.956: INFO: Got endpoints: latency-svc-292br [677.141427ms] Apr 25 00:15:57.974: INFO: Created: latency-svc-kc4rk Apr 25 00:15:57.987: INFO: Got endpoints: latency-svc-kc4rk [635.317275ms] Apr 25 00:15:58.005: INFO: Created: latency-svc-b8ltr Apr 25 00:15:58.036: INFO: Got endpoints: latency-svc-b8ltr [637.001475ms] Apr 25 00:15:58.083: INFO: Created: latency-svc-h8hpb Apr 25 00:15:58.107: INFO: Created: latency-svc-fc2vc Apr 25 00:15:58.107: INFO: Got endpoints: latency-svc-h8hpb [671.76642ms] Apr 25 00:15:58.123: INFO: Got endpoints: latency-svc-fc2vc [616.279309ms] Apr 25 00:15:58.149: INFO: Created: latency-svc-c2t2b Apr 25 00:15:58.167: INFO: Got endpoints: latency-svc-c2t2b [642.915839ms] Apr 25 00:15:58.221: INFO: Created: latency-svc-xddsg Apr 25 00:15:58.261: INFO: Got endpoints: latency-svc-xddsg [701.406306ms] Apr 25 00:15:58.281: INFO: Created: latency-svc-brdfd Apr 25 00:15:58.352: INFO: Got endpoints: latency-svc-brdfd [712.33692ms] Apr 25 00:15:58.370: INFO: Created: latency-svc-qv5xf Apr 25 00:15:58.387: INFO: Got endpoints: latency-svc-qv5xf [737.302859ms] Apr 25 00:15:58.412: INFO: Created: latency-svc-smljl Apr 25 00:15:58.423: INFO: Got endpoints: latency-svc-smljl [728.749758ms] Apr 25 00:15:58.448: INFO: Created: latency-svc-w48sw Apr 25 00:15:58.502: INFO: Got endpoints: latency-svc-w48sw [774.49628ms] Apr 25 00:15:58.515: INFO: Created: latency-svc-b9qwp Apr 25 00:15:58.525: INFO: Got endpoints: latency-svc-b9qwp [713.970965ms] Apr 25 00:15:58.539: INFO: Created: latency-svc-g9w4v Apr 25 00:15:58.550: INFO: Got endpoints: latency-svc-g9w4v [700.633159ms] Apr 25 00:15:58.562: INFO: Created: latency-svc-qr44m Apr 25 00:15:58.580: INFO: Got endpoints: latency-svc-qr44m [700.889146ms] Apr 25 00:15:58.598: INFO: Created: latency-svc-l2kxr Apr 25 00:15:58.627: INFO: Got endpoints: latency-svc-l2kxr [682.092352ms] Apr 25 00:15:58.640: INFO: Created: latency-svc-jjs5l Apr 25 00:15:58.658: INFO: Got endpoints: latency-svc-jjs5l [701.259993ms] Apr 25 00:15:58.676: INFO: Created: latency-svc-27528 Apr 25 00:15:58.694: INFO: Got endpoints: latency-svc-27528 [706.097316ms] Apr 25 00:15:58.713: INFO: Created: latency-svc-pznqr Apr 25 00:15:58.771: INFO: Got endpoints: latency-svc-pznqr [734.818727ms] Apr 25 00:15:58.790: INFO: Created: latency-svc-t6vmp Apr 25 00:15:58.807: INFO: Got endpoints: latency-svc-t6vmp [699.778522ms] Apr 25 00:15:58.826: INFO: Created: latency-svc-sqgqf Apr 25 00:15:58.836: INFO: Got endpoints: latency-svc-sqgqf [712.982799ms] Apr 25 00:15:58.862: INFO: Created: latency-svc-r9q7k Apr 25 00:15:58.909: INFO: Got endpoints: latency-svc-r9q7k [742.536035ms] Apr 25 00:15:58.942: INFO: Created: latency-svc-nqqxf Apr 25 00:15:58.956: INFO: Got endpoints: latency-svc-nqqxf [694.708975ms] Apr 25 00:15:58.977: INFO: Created: latency-svc-gz6qw Apr 25 00:15:58.992: INFO: Got endpoints: latency-svc-gz6qw [639.579742ms] Apr 25 00:15:59.065: INFO: Created: latency-svc-9bcgk Apr 25 00:15:59.097: INFO: Created: latency-svc-psfpz Apr 25 00:15:59.097: INFO: Got endpoints: latency-svc-9bcgk [710.533273ms] Apr 25 00:15:59.127: INFO: Got endpoints: latency-svc-psfpz [703.640504ms] Apr 25 00:15:59.145: INFO: Created: latency-svc-bmrc9 Apr 25 00:15:59.155: INFO: Got endpoints: latency-svc-bmrc9 [652.715349ms] Apr 25 00:15:59.155: INFO: Latencies: [34.103851ms 75.241102ms 111.240815ms 158.720962ms 207.249702ms 220.424911ms 250.058565ms 280.024193ms 341.333104ms 351.843596ms 388.139416ms 417.860337ms 452.010809ms 481.892806ms 511.771117ms 519.499316ms 522.197384ms 533.07192ms 534.807911ms 538.78532ms 539.436061ms 539.852715ms 558.10577ms 570.023691ms 575.317977ms 593.068503ms 593.219519ms 599.11693ms 599.686657ms 599.874146ms 599.955655ms 603.056198ms 603.391998ms 604.46694ms 609.73047ms 616.279309ms 616.366769ms 620.656159ms 620.82082ms 623.567267ms 626.376531ms 635.317275ms 637.001475ms 639.579742ms 642.343786ms 642.915839ms 652.542445ms 652.715349ms 656.105745ms 659.097494ms 661.938139ms 669.462041ms 671.302615ms 671.76642ms 673.082458ms 673.087368ms 675.917397ms 676.894778ms 677.141427ms 678.73167ms 682.092352ms 683.007244ms 686.863342ms 688.728469ms 688.955274ms 691.410826ms 692.63575ms 694.448542ms 694.708975ms 696.090214ms 699.778522ms 699.865185ms 700.44658ms 700.633159ms 700.889146ms 701.10252ms 701.259993ms 701.33634ms 701.406306ms 701.873849ms 702.155633ms 703.640504ms 704.254844ms 706.097316ms 706.254043ms 706.664204ms 710.010344ms 710.533273ms 712.292117ms 712.33692ms 712.736553ms 712.946845ms 712.982799ms 713.970965ms 714.084722ms 723.128006ms 723.622401ms 724.631865ms 724.824829ms 728.749758ms 731.672296ms 734.818727ms 736.747644ms 737.302859ms 741.4314ms 741.578538ms 742.389421ms 742.536035ms 743.255332ms 766.342863ms 773.114404ms 774.49628ms 790.917752ms 799.019339ms 803.006497ms 814.873231ms 823.673829ms 833.354797ms 840.560074ms 844.387022ms 859.502334ms 881.995048ms 895.444923ms 910.341677ms 960.35144ms 965.663068ms 977.470324ms 988.766673ms 1.001211529s 1.018867753s 1.021576481s 1.051503868s 1.0607628s 1.061218672s 1.064205121s 1.065065173s 1.066109125s 1.066295014s 1.068035649s 1.070682033s 1.111250216s 1.162142395s 1.18965897s 1.232552198s 1.275257928s 1.275668992s 1.290136371s 1.298657684s 1.301340133s 1.312109336s 1.316074483s 1.318987784s 1.321937649s 1.335424011s 1.336505429s 1.341729543s 1.347485404s 1.347542358s 1.349860707s 1.359087256s 1.361655877s 1.367225857s 1.383804044s 1.387492572s 1.394515011s 1.401126399s 1.407259202s 1.424452469s 1.451307929s 1.519741038s 1.522004579s 1.533542009s 1.618464138s 1.655753087s 1.657097152s 1.783441456s 1.791987519s 1.793166483s 1.89761547s 1.979246739s 2.039748673s 2.051635026s 2.105354656s 2.192488823s 2.197558643s 2.224104834s 2.458945531s 2.477581247s 2.501135115s 2.563043666s 2.803069522s 2.865567822s 2.913593858s 3.013855792s 3.026514319s 3.046578069s 3.084202767s 3.088330257s 3.12164158s 3.156567769s] Apr 25 00:15:59.155: INFO: 50 %ile: 731.672296ms Apr 25 00:15:59.155: INFO: 90 %ile: 2.039748673s Apr 25 00:15:59.155: INFO: 99 %ile: 3.12164158s Apr 25 00:15:59.155: INFO: Total sample count: 200 [AfterEach] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:15:59.155: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svc-latency-9871" for this suite. • [SLOW TEST:17.455 seconds] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Service endpoints latency should not be very high [Conformance]","total":275,"completed":133,"skipped":2357,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:15:59.216: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:91 Apr 25 00:15:59.289: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 25 00:15:59.333: INFO: Waiting for terminating namespaces to be deleted... Apr 25 00:15:59.334: INFO: Logging pods the kubelet thinks is on node latest-worker before test Apr 25 00:15:59.338: INFO: kindnet-vnjgh from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:15:59.338: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:15:59.338: INFO: kube-proxy-s9v6p from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:15:59.338: INFO: Container kube-proxy ready: true, restart count 0 Apr 25 00:15:59.338: INFO: svc-latency-rc-hwlbg from svc-latency-9871 started at 2020-04-25 00:15:41 +0000 UTC (1 container statuses recorded) Apr 25 00:15:59.338: INFO: Container svc-latency-rc ready: true, restart count 0 Apr 25 00:15:59.338: INFO: Logging pods the kubelet thinks is on node latest-worker2 before test Apr 25 00:15:59.341: INFO: kindnet-zq6gp from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:15:59.341: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:15:59.341: INFO: kube-proxy-c5xlk from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:15:59.341: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-add6d804-4723-42d3-837f-57e586cc8561 95 STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 127.0.0.1 on the node which pod4 resides and expect not scheduled STEP: removing the label kubernetes.io/e2e-add6d804-4723-42d3-837f-57e586cc8561 off the node latest-worker STEP: verifying the node doesn't have the label kubernetes.io/e2e-add6d804-4723-42d3-837f-57e586cc8561 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:21:07.613: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-5910" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:82 • [SLOW TEST:308.404 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","total":275,"completed":134,"skipped":2393,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:21:07.621: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:84 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:99 STEP: Creating service test in namespace statefulset-1084 [It] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Looking for a node to schedule stateful set and pod STEP: Creating pod with conflicting port in namespace statefulset-1084 STEP: Creating statefulset with conflicting port in namespace statefulset-1084 STEP: Waiting until pod test-pod will start running in namespace statefulset-1084 STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-1084 Apr 25 00:21:11.747: INFO: Observed stateful pod in namespace: statefulset-1084, name: ss-0, uid: 03a9e201-3273-4ca2-a8c1-0e04bc5d0462, status phase: Pending. Waiting for statefulset controller to delete. Apr 25 00:21:12.972: INFO: Observed stateful pod in namespace: statefulset-1084, name: ss-0, uid: 03a9e201-3273-4ca2-a8c1-0e04bc5d0462, status phase: Failed. Waiting for statefulset controller to delete. Apr 25 00:21:12.980: INFO: Observed stateful pod in namespace: statefulset-1084, name: ss-0, uid: 03a9e201-3273-4ca2-a8c1-0e04bc5d0462, status phase: Failed. Waiting for statefulset controller to delete. Apr 25 00:21:12.992: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-1084 STEP: Removing pod with conflicting port in namespace statefulset-1084 STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-1084 and will be in running state [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:110 Apr 25 00:21:25.806: INFO: Deleting all statefulset in ns statefulset-1084 Apr 25 00:21:25.810: INFO: Scaling statefulset ss to 0 Apr 25 00:21:35.828: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:21:35.831: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:21:35.843: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-1084" for this suite. • [SLOW TEST:28.229 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","total":275,"completed":135,"skipped":2405,"failed":0} SSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:21:35.851: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for deployment deletion to see if the garbage collector mistakenly deletes the rs STEP: Gathering metrics W0425 00:21:36.946707 8 metrics_grabber.go:84] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 25 00:21:36.946: INFO: For apiserver_request_total: For apiserver_request_latency_seconds: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:21:36.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-8799" for this suite. •{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","total":275,"completed":136,"skipped":2409,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:21:36.953: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: validating cluster-info Apr 25 00:21:37.030: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config cluster-info' Apr 25 00:21:37.135: INFO: stderr: "" Apr 25 00:21:37.135: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:32771\x1b[0m\n\x1b[0;32mKubeDNS\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:32771/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:21:37.135: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5624" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance]","total":275,"completed":137,"skipped":2426,"failed":0} SSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:21:37.142: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:21:37.190: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:21:38.219: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-2256" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance]","total":275,"completed":138,"skipped":2435,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:21:38.227: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating Agnhost RC Apr 25 00:21:38.334: INFO: namespace kubectl-482 Apr 25 00:21:38.335: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-482' Apr 25 00:21:38.786: INFO: stderr: "" Apr 25 00:21:38.786: INFO: stdout: "replicationcontroller/agnhost-master created\n" STEP: Waiting for Agnhost master to start. Apr 25 00:21:39.800: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:21:39.800: INFO: Found 0 / 1 Apr 25 00:21:40.855: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:21:40.855: INFO: Found 0 / 1 Apr 25 00:21:41.801: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:21:41.801: INFO: Found 1 / 1 Apr 25 00:21:41.801: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Apr 25 00:21:41.804: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:21:41.804: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Apr 25 00:21:41.804: INFO: wait on agnhost-master startup in kubectl-482 Apr 25 00:21:41.804: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config logs agnhost-master-tdcm7 agnhost-master --namespace=kubectl-482' Apr 25 00:21:41.922: INFO: stderr: "" Apr 25 00:21:41.922: INFO: stdout: "Paused\n" STEP: exposing RC Apr 25 00:21:41.922: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config expose rc agnhost-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-482' Apr 25 00:21:42.066: INFO: stderr: "" Apr 25 00:21:42.066: INFO: stdout: "service/rm2 exposed\n" Apr 25 00:21:42.076: INFO: Service rm2 in namespace kubectl-482 found. STEP: exposing service Apr 25 00:21:44.083: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-482' Apr 25 00:21:44.210: INFO: stderr: "" Apr 25 00:21:44.210: INFO: stdout: "service/rm3 exposed\n" Apr 25 00:21:44.238: INFO: Service rm3 in namespace kubectl-482 found. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:21:46.247: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-482" for this suite. • [SLOW TEST:8.029 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl expose /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1119 should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance]","total":275,"completed":139,"skipped":2451,"failed":0} SSS ------------------------------ [sig-network] DNS should provide DNS for pods for Subdomain [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:21:46.257: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for pods for Subdomain [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6539.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-6539.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6539.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:21:52.365: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.369: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.372: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.375: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.385: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.388: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.391: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.394: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:52.400: INFO: Lookups using dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local] Apr 25 00:21:57.404: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.408: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.410: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.412: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.420: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.423: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.426: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.428: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:21:57.433: INFO: Lookups using dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local] Apr 25 00:22:02.406: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.410: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.413: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.416: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.425: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.428: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.430: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.433: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:02.440: INFO: Lookups using dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local] Apr 25 00:22:07.405: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.409: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.412: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.415: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.424: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.427: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.430: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.433: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:07.437: INFO: Lookups using dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local] Apr 25 00:22:12.405: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.409: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.413: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.416: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.427: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.430: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.433: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.436: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:12.443: INFO: Lookups using dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local] Apr 25 00:22:17.406: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.408: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.410: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.412: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.419: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.422: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.424: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.426: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local from pod dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de: the server could not find the requested resource (get pods dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de) Apr 25 00:22:17.430: INFO: Lookups using dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6539.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6539.svc.cluster.local jessie_udp@dns-test-service-2.dns-6539.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6539.svc.cluster.local] Apr 25 00:22:22.456: INFO: DNS probes using dns-6539/dns-test-02e5d448-06dd-44ad-ac74-14ae289b29de succeeded STEP: deleting the pod STEP: deleting the test headless service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:22:22.539: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-6539" for this suite. • [SLOW TEST:36.516 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for pods for Subdomain [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":275,"completed":140,"skipped":2454,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:22:22.774: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:126 STEP: Setting up server cert STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication STEP: Deploying the custom resource conversion webhook pod STEP: Wait for the deployment to be ready Apr 25 00:22:23.604: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set Apr 25 00:22:25.615: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370943, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370943, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370943, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723370943, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-54c8b67c75\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:22:28.641: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 [It] should be able to convert a non homogeneous list of CRs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:22:28.645: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating a v1 custom resource STEP: Create a v2 custom resource STEP: List CRs in v1 STEP: List CRs in v2 [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:22:29.900: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-webhook-2715" for this suite. [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:137 • [SLOW TEST:7.240 seconds] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to convert a non homogeneous list of CRs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","total":275,"completed":141,"skipped":2462,"failed":0} SSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:22:30.013: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-volume-65021303-41a9-4e93-82fc-cc34304b2563 STEP: Creating a pod to test consume configMaps Apr 25 00:22:30.140: INFO: Waiting up to 5m0s for pod "pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072" in namespace "configmap-2619" to be "Succeeded or Failed" Apr 25 00:22:30.143: INFO: Pod "pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072": Phase="Pending", Reason="", readiness=false. Elapsed: 3.545841ms Apr 25 00:22:32.152: INFO: Pod "pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012521292s Apr 25 00:22:34.156: INFO: Pod "pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016262153s STEP: Saw pod success Apr 25 00:22:34.156: INFO: Pod "pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072" satisfied condition "Succeeded or Failed" Apr 25 00:22:34.159: INFO: Trying to get logs from node latest-worker2 pod pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072 container configmap-volume-test: STEP: delete the pod Apr 25 00:22:34.174: INFO: Waiting for pod pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072 to disappear Apr 25 00:22:34.179: INFO: Pod pod-configmaps-c44fcc1d-397e-4142-b1be-7378c52e6072 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:22:34.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2619" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":142,"skipped":2469,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:22:34.202: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] updates the published spec when one version gets renamed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: set up a multi version CRD Apr 25 00:22:34.284: INFO: >>> kubeConfig: /root/.kube/config STEP: rename a version STEP: check the new version name is served STEP: check the old version name is removed STEP: check the other version is not changed [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:22:50.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-1121" for this suite. • [SLOW TEST:15.855 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 updates the published spec when one version gets renamed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","total":275,"completed":143,"skipped":2505,"failed":0} SSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:22:50.057: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:84 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:99 STEP: Creating service test in namespace statefulset-734 [It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating stateful set ss in namespace statefulset-734 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-734 Apr 25 00:22:50.197: INFO: Found 0 stateful pods, waiting for 1 Apr 25 00:23:00.202: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod Apr 25 00:23:00.206: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-734 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:23:03.109: INFO: stderr: "I0425 00:23:02.995174 1375 log.go:172] (0xc0008a6840) (0xc0008e21e0) Create stream\nI0425 00:23:02.995224 1375 log.go:172] (0xc0008a6840) (0xc0008e21e0) Stream added, broadcasting: 1\nI0425 00:23:02.999021 1375 log.go:172] (0xc0008a6840) Reply frame received for 1\nI0425 00:23:02.999066 1375 log.go:172] (0xc0008a6840) (0xc0008aa0a0) Create stream\nI0425 00:23:02.999076 1375 log.go:172] (0xc0008a6840) (0xc0008aa0a0) Stream added, broadcasting: 3\nI0425 00:23:03.000331 1375 log.go:172] (0xc0008a6840) Reply frame received for 3\nI0425 00:23:03.000375 1375 log.go:172] (0xc0008a6840) (0xc0008a00a0) Create stream\nI0425 00:23:03.000388 1375 log.go:172] (0xc0008a6840) (0xc0008a00a0) Stream added, broadcasting: 5\nI0425 00:23:03.001598 1375 log.go:172] (0xc0008a6840) Reply frame received for 5\nI0425 00:23:03.081679 1375 log.go:172] (0xc0008a6840) Data frame received for 5\nI0425 00:23:03.081714 1375 log.go:172] (0xc0008a00a0) (5) Data frame handling\nI0425 00:23:03.081744 1375 log.go:172] (0xc0008a00a0) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:23:03.099770 1375 log.go:172] (0xc0008a6840) Data frame received for 3\nI0425 00:23:03.099795 1375 log.go:172] (0xc0008aa0a0) (3) Data frame handling\nI0425 00:23:03.099808 1375 log.go:172] (0xc0008aa0a0) (3) Data frame sent\nI0425 00:23:03.099915 1375 log.go:172] (0xc0008a6840) Data frame received for 5\nI0425 00:23:03.099925 1375 log.go:172] (0xc0008a00a0) (5) Data frame handling\nI0425 00:23:03.100636 1375 log.go:172] (0xc0008a6840) Data frame received for 3\nI0425 00:23:03.100655 1375 log.go:172] (0xc0008aa0a0) (3) Data frame handling\nI0425 00:23:03.103923 1375 log.go:172] (0xc0008a6840) Data frame received for 1\nI0425 00:23:03.103947 1375 log.go:172] (0xc0008e21e0) (1) Data frame handling\nI0425 00:23:03.103957 1375 log.go:172] (0xc0008e21e0) (1) Data frame sent\nI0425 00:23:03.103966 1375 log.go:172] (0xc0008a6840) (0xc0008e21e0) Stream removed, broadcasting: 1\nI0425 00:23:03.103975 1375 log.go:172] (0xc0008a6840) Go away received\nI0425 00:23:03.104500 1375 log.go:172] (0xc0008a6840) (0xc0008e21e0) Stream removed, broadcasting: 1\nI0425 00:23:03.104528 1375 log.go:172] (0xc0008a6840) (0xc0008aa0a0) Stream removed, broadcasting: 3\nI0425 00:23:03.104548 1375 log.go:172] (0xc0008a6840) (0xc0008a00a0) Stream removed, broadcasting: 5\n" Apr 25 00:23:03.109: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:23:03.109: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:23:03.112: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Apr 25 00:23:13.118: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:23:13.118: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:23:13.148: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:13.148: INFO: ss-0 latest-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:03 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:03 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC }] Apr 25 00:23:13.148: INFO: Apr 25 00:23:13.148: INFO: StatefulSet ss has not reached scale 3, at 1 Apr 25 00:23:14.153: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.979897022s Apr 25 00:23:15.158: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.974947137s Apr 25 00:23:16.175: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.970201857s Apr 25 00:23:17.179: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.953645901s Apr 25 00:23:18.184: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.949146576s Apr 25 00:23:19.190: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.94356226s Apr 25 00:23:20.195: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.938445672s Apr 25 00:23:21.200: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.93274194s Apr 25 00:23:22.205: INFO: Verifying statefulset ss doesn't scale past 3 for another 928.026937ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-734 Apr 25 00:23:23.210: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-734 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:23:23.468: INFO: stderr: "I0425 00:23:23.352196 1409 log.go:172] (0xc000910000) (0xc00082d220) Create stream\nI0425 00:23:23.352264 1409 log.go:172] (0xc000910000) (0xc00082d220) Stream added, broadcasting: 1\nI0425 00:23:23.354922 1409 log.go:172] (0xc000910000) Reply frame received for 1\nI0425 00:23:23.354979 1409 log.go:172] (0xc000910000) (0xc000ba4000) Create stream\nI0425 00:23:23.354994 1409 log.go:172] (0xc000910000) (0xc000ba4000) Stream added, broadcasting: 3\nI0425 00:23:23.356086 1409 log.go:172] (0xc000910000) Reply frame received for 3\nI0425 00:23:23.356116 1409 log.go:172] (0xc000910000) (0xc000ba40a0) Create stream\nI0425 00:23:23.356135 1409 log.go:172] (0xc000910000) (0xc000ba40a0) Stream added, broadcasting: 5\nI0425 00:23:23.357427 1409 log.go:172] (0xc000910000) Reply frame received for 5\nI0425 00:23:23.462549 1409 log.go:172] (0xc000910000) Data frame received for 3\nI0425 00:23:23.462601 1409 log.go:172] (0xc000ba4000) (3) Data frame handling\nI0425 00:23:23.462611 1409 log.go:172] (0xc000ba4000) (3) Data frame sent\nI0425 00:23:23.462619 1409 log.go:172] (0xc000910000) Data frame received for 3\nI0425 00:23:23.462625 1409 log.go:172] (0xc000ba4000) (3) Data frame handling\nI0425 00:23:23.462649 1409 log.go:172] (0xc000910000) Data frame received for 5\nI0425 00:23:23.462656 1409 log.go:172] (0xc000ba40a0) (5) Data frame handling\nI0425 00:23:23.462665 1409 log.go:172] (0xc000ba40a0) (5) Data frame sent\nI0425 00:23:23.462678 1409 log.go:172] (0xc000910000) Data frame received for 5\nI0425 00:23:23.462691 1409 log.go:172] (0xc000ba40a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI0425 00:23:23.464288 1409 log.go:172] (0xc000910000) Data frame received for 1\nI0425 00:23:23.464322 1409 log.go:172] (0xc00082d220) (1) Data frame handling\nI0425 00:23:23.464334 1409 log.go:172] (0xc00082d220) (1) Data frame sent\nI0425 00:23:23.464347 1409 log.go:172] (0xc000910000) (0xc00082d220) Stream removed, broadcasting: 1\nI0425 00:23:23.464384 1409 log.go:172] (0xc000910000) Go away received\nI0425 00:23:23.464670 1409 log.go:172] (0xc000910000) (0xc00082d220) Stream removed, broadcasting: 1\nI0425 00:23:23.464684 1409 log.go:172] (0xc000910000) (0xc000ba4000) Stream removed, broadcasting: 3\nI0425 00:23:23.464690 1409 log.go:172] (0xc000910000) (0xc000ba40a0) Stream removed, broadcasting: 5\n" Apr 25 00:23:23.468: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:23:23.468: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:23:23.469: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-734 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:23:23.706: INFO: stderr: "I0425 00:23:23.628626 1431 log.go:172] (0xc000aacf20) (0xc000ab6640) Create stream\nI0425 00:23:23.628692 1431 log.go:172] (0xc000aacf20) (0xc000ab6640) Stream added, broadcasting: 1\nI0425 00:23:23.631264 1431 log.go:172] (0xc000aacf20) Reply frame received for 1\nI0425 00:23:23.631297 1431 log.go:172] (0xc000aacf20) (0xc000524aa0) Create stream\nI0425 00:23:23.631310 1431 log.go:172] (0xc000aacf20) (0xc000524aa0) Stream added, broadcasting: 3\nI0425 00:23:23.632386 1431 log.go:172] (0xc000aacf20) Reply frame received for 3\nI0425 00:23:23.632406 1431 log.go:172] (0xc000aacf20) (0xc000ab66e0) Create stream\nI0425 00:23:23.632414 1431 log.go:172] (0xc000aacf20) (0xc000ab66e0) Stream added, broadcasting: 5\nI0425 00:23:23.633586 1431 log.go:172] (0xc000aacf20) Reply frame received for 5\nI0425 00:23:23.698739 1431 log.go:172] (0xc000aacf20) Data frame received for 3\nI0425 00:23:23.698796 1431 log.go:172] (0xc000524aa0) (3) Data frame handling\nI0425 00:23:23.698821 1431 log.go:172] (0xc000524aa0) (3) Data frame sent\nI0425 00:23:23.698843 1431 log.go:172] (0xc000aacf20) Data frame received for 3\nI0425 00:23:23.698855 1431 log.go:172] (0xc000524aa0) (3) Data frame handling\nI0425 00:23:23.698892 1431 log.go:172] (0xc000aacf20) Data frame received for 5\nI0425 00:23:23.698921 1431 log.go:172] (0xc000ab66e0) (5) Data frame handling\nI0425 00:23:23.698954 1431 log.go:172] (0xc000ab66e0) (5) Data frame sent\nI0425 00:23:23.698969 1431 log.go:172] (0xc000aacf20) Data frame received for 5\nI0425 00:23:23.698982 1431 log.go:172] (0xc000ab66e0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI0425 00:23:23.700224 1431 log.go:172] (0xc000aacf20) Data frame received for 1\nI0425 00:23:23.700247 1431 log.go:172] (0xc000ab6640) (1) Data frame handling\nI0425 00:23:23.700270 1431 log.go:172] (0xc000ab6640) (1) Data frame sent\nI0425 00:23:23.700323 1431 log.go:172] (0xc000aacf20) (0xc000ab6640) Stream removed, broadcasting: 1\nI0425 00:23:23.700353 1431 log.go:172] (0xc000aacf20) Go away received\nI0425 00:23:23.700762 1431 log.go:172] (0xc000aacf20) (0xc000ab6640) Stream removed, broadcasting: 1\nI0425 00:23:23.700785 1431 log.go:172] (0xc000aacf20) (0xc000524aa0) Stream removed, broadcasting: 3\nI0425 00:23:23.700800 1431 log.go:172] (0xc000aacf20) (0xc000ab66e0) Stream removed, broadcasting: 5\n" Apr 25 00:23:23.706: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:23:23.706: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:23:23.706: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-734 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:23:23.922: INFO: stderr: "I0425 00:23:23.828240 1452 log.go:172] (0xc000aa5340) (0xc000ac25a0) Create stream\nI0425 00:23:23.828337 1452 log.go:172] (0xc000aa5340) (0xc000ac25a0) Stream added, broadcasting: 1\nI0425 00:23:23.831689 1452 log.go:172] (0xc000aa5340) Reply frame received for 1\nI0425 00:23:23.831728 1452 log.go:172] (0xc000aa5340) (0xc000ac2640) Create stream\nI0425 00:23:23.831740 1452 log.go:172] (0xc000aa5340) (0xc000ac2640) Stream added, broadcasting: 3\nI0425 00:23:23.832844 1452 log.go:172] (0xc000aa5340) Reply frame received for 3\nI0425 00:23:23.832884 1452 log.go:172] (0xc000aa5340) (0xc000ab0280) Create stream\nI0425 00:23:23.832896 1452 log.go:172] (0xc000aa5340) (0xc000ab0280) Stream added, broadcasting: 5\nI0425 00:23:23.834062 1452 log.go:172] (0xc000aa5340) Reply frame received for 5\nI0425 00:23:23.914134 1452 log.go:172] (0xc000aa5340) Data frame received for 5\nI0425 00:23:23.914165 1452 log.go:172] (0xc000ab0280) (5) Data frame handling\nI0425 00:23:23.914192 1452 log.go:172] (0xc000ab0280) (5) Data frame sent\nI0425 00:23:23.914205 1452 log.go:172] (0xc000aa5340) Data frame received for 5\nI0425 00:23:23.914214 1452 log.go:172] (0xc000ab0280) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI0425 00:23:23.914272 1452 log.go:172] (0xc000aa5340) Data frame received for 3\nI0425 00:23:23.914304 1452 log.go:172] (0xc000ac2640) (3) Data frame handling\nI0425 00:23:23.914317 1452 log.go:172] (0xc000ac2640) (3) Data frame sent\nI0425 00:23:23.914326 1452 log.go:172] (0xc000aa5340) Data frame received for 3\nI0425 00:23:23.914335 1452 log.go:172] (0xc000ac2640) (3) Data frame handling\nI0425 00:23:23.916343 1452 log.go:172] (0xc000aa5340) Data frame received for 1\nI0425 00:23:23.916360 1452 log.go:172] (0xc000ac25a0) (1) Data frame handling\nI0425 00:23:23.916370 1452 log.go:172] (0xc000ac25a0) (1) Data frame sent\nI0425 00:23:23.916382 1452 log.go:172] (0xc000aa5340) (0xc000ac25a0) Stream removed, broadcasting: 1\nI0425 00:23:23.916398 1452 log.go:172] (0xc000aa5340) Go away received\nI0425 00:23:23.916780 1452 log.go:172] (0xc000aa5340) (0xc000ac25a0) Stream removed, broadcasting: 1\nI0425 00:23:23.916805 1452 log.go:172] (0xc000aa5340) (0xc000ac2640) Stream removed, broadcasting: 3\nI0425 00:23:23.916818 1452 log.go:172] (0xc000aa5340) (0xc000ab0280) Stream removed, broadcasting: 5\n" Apr 25 00:23:23.922: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:23:23.922: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:23:23.926: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:23:23.926: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:23:23.926: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Scale down will not halt with unhealthy stateful pod Apr 25 00:23:23.929: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-734 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:23:24.159: INFO: stderr: "I0425 00:23:24.077702 1474 log.go:172] (0xc0003cda20) (0xc000a8a1e0) Create stream\nI0425 00:23:24.077779 1474 log.go:172] (0xc0003cda20) (0xc000a8a1e0) Stream added, broadcasting: 1\nI0425 00:23:24.081455 1474 log.go:172] (0xc0003cda20) Reply frame received for 1\nI0425 00:23:24.081514 1474 log.go:172] (0xc0003cda20) (0xc0006cd2c0) Create stream\nI0425 00:23:24.081521 1474 log.go:172] (0xc0003cda20) (0xc0006cd2c0) Stream added, broadcasting: 3\nI0425 00:23:24.082469 1474 log.go:172] (0xc0003cda20) Reply frame received for 3\nI0425 00:23:24.082506 1474 log.go:172] (0xc0003cda20) (0xc000a8a280) Create stream\nI0425 00:23:24.082518 1474 log.go:172] (0xc0003cda20) (0xc000a8a280) Stream added, broadcasting: 5\nI0425 00:23:24.083627 1474 log.go:172] (0xc0003cda20) Reply frame received for 5\nI0425 00:23:24.152544 1474 log.go:172] (0xc0003cda20) Data frame received for 3\nI0425 00:23:24.152649 1474 log.go:172] (0xc0006cd2c0) (3) Data frame handling\nI0425 00:23:24.152672 1474 log.go:172] (0xc0006cd2c0) (3) Data frame sent\nI0425 00:23:24.152684 1474 log.go:172] (0xc0003cda20) Data frame received for 3\nI0425 00:23:24.152692 1474 log.go:172] (0xc0006cd2c0) (3) Data frame handling\nI0425 00:23:24.152723 1474 log.go:172] (0xc0003cda20) Data frame received for 5\nI0425 00:23:24.152734 1474 log.go:172] (0xc000a8a280) (5) Data frame handling\nI0425 00:23:24.152745 1474 log.go:172] (0xc000a8a280) (5) Data frame sent\nI0425 00:23:24.152755 1474 log.go:172] (0xc0003cda20) Data frame received for 5\nI0425 00:23:24.152763 1474 log.go:172] (0xc000a8a280) (5) Data frame handling\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:23:24.154630 1474 log.go:172] (0xc0003cda20) Data frame received for 1\nI0425 00:23:24.154673 1474 log.go:172] (0xc000a8a1e0) (1) Data frame handling\nI0425 00:23:24.154704 1474 log.go:172] (0xc000a8a1e0) (1) Data frame sent\nI0425 00:23:24.154733 1474 log.go:172] (0xc0003cda20) (0xc000a8a1e0) Stream removed, broadcasting: 1\nI0425 00:23:24.154770 1474 log.go:172] (0xc0003cda20) Go away received\nI0425 00:23:24.155048 1474 log.go:172] (0xc0003cda20) (0xc000a8a1e0) Stream removed, broadcasting: 1\nI0425 00:23:24.155066 1474 log.go:172] (0xc0003cda20) (0xc0006cd2c0) Stream removed, broadcasting: 3\nI0425 00:23:24.155075 1474 log.go:172] (0xc0003cda20) (0xc000a8a280) Stream removed, broadcasting: 5\n" Apr 25 00:23:24.159: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:23:24.159: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:23:24.159: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-734 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:23:24.390: INFO: stderr: "I0425 00:23:24.283625 1496 log.go:172] (0xc000a0ec60) (0xc0009f0320) Create stream\nI0425 00:23:24.283670 1496 log.go:172] (0xc000a0ec60) (0xc0009f0320) Stream added, broadcasting: 1\nI0425 00:23:24.286915 1496 log.go:172] (0xc000a0ec60) Reply frame received for 1\nI0425 00:23:24.286949 1496 log.go:172] (0xc000a0ec60) (0xc000a76000) Create stream\nI0425 00:23:24.286959 1496 log.go:172] (0xc000a0ec60) (0xc000a76000) Stream added, broadcasting: 3\nI0425 00:23:24.287530 1496 log.go:172] (0xc000a0ec60) Reply frame received for 3\nI0425 00:23:24.287558 1496 log.go:172] (0xc000a0ec60) (0xc0009f0000) Create stream\nI0425 00:23:24.287575 1496 log.go:172] (0xc000a0ec60) (0xc0009f0000) Stream added, broadcasting: 5\nI0425 00:23:24.288202 1496 log.go:172] (0xc000a0ec60) Reply frame received for 5\nI0425 00:23:24.351568 1496 log.go:172] (0xc000a0ec60) Data frame received for 5\nI0425 00:23:24.351595 1496 log.go:172] (0xc0009f0000) (5) Data frame handling\nI0425 00:23:24.351611 1496 log.go:172] (0xc0009f0000) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:23:24.382292 1496 log.go:172] (0xc000a0ec60) Data frame received for 3\nI0425 00:23:24.382322 1496 log.go:172] (0xc000a76000) (3) Data frame handling\nI0425 00:23:24.382341 1496 log.go:172] (0xc000a76000) (3) Data frame sent\nI0425 00:23:24.382468 1496 log.go:172] (0xc000a0ec60) Data frame received for 3\nI0425 00:23:24.382497 1496 log.go:172] (0xc000a76000) (3) Data frame handling\nI0425 00:23:24.382536 1496 log.go:172] (0xc000a0ec60) Data frame received for 5\nI0425 00:23:24.382549 1496 log.go:172] (0xc0009f0000) (5) Data frame handling\nI0425 00:23:24.384231 1496 log.go:172] (0xc000a0ec60) Data frame received for 1\nI0425 00:23:24.384247 1496 log.go:172] (0xc0009f0320) (1) Data frame handling\nI0425 00:23:24.384257 1496 log.go:172] (0xc0009f0320) (1) Data frame sent\nI0425 00:23:24.384266 1496 log.go:172] (0xc000a0ec60) (0xc0009f0320) Stream removed, broadcasting: 1\nI0425 00:23:24.384457 1496 log.go:172] (0xc000a0ec60) Go away received\nI0425 00:23:24.384558 1496 log.go:172] (0xc000a0ec60) (0xc0009f0320) Stream removed, broadcasting: 1\nI0425 00:23:24.384578 1496 log.go:172] (0xc000a0ec60) (0xc000a76000) Stream removed, broadcasting: 3\nI0425 00:23:24.384586 1496 log.go:172] (0xc000a0ec60) (0xc0009f0000) Stream removed, broadcasting: 5\n" Apr 25 00:23:24.390: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:23:24.390: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:23:24.390: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-734 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:23:24.634: INFO: stderr: "I0425 00:23:24.514183 1517 log.go:172] (0xc000a7f1e0) (0xc0009e2500) Create stream\nI0425 00:23:24.514261 1517 log.go:172] (0xc000a7f1e0) (0xc0009e2500) Stream added, broadcasting: 1\nI0425 00:23:24.519992 1517 log.go:172] (0xc000a7f1e0) Reply frame received for 1\nI0425 00:23:24.520031 1517 log.go:172] (0xc000a7f1e0) (0xc000405720) Create stream\nI0425 00:23:24.520044 1517 log.go:172] (0xc000a7f1e0) (0xc000405720) Stream added, broadcasting: 3\nI0425 00:23:24.520948 1517 log.go:172] (0xc000a7f1e0) Reply frame received for 3\nI0425 00:23:24.520992 1517 log.go:172] (0xc000a7f1e0) (0xc0005b1900) Create stream\nI0425 00:23:24.521009 1517 log.go:172] (0xc000a7f1e0) (0xc0005b1900) Stream added, broadcasting: 5\nI0425 00:23:24.522001 1517 log.go:172] (0xc000a7f1e0) Reply frame received for 5\nI0425 00:23:24.575155 1517 log.go:172] (0xc000a7f1e0) Data frame received for 5\nI0425 00:23:24.575205 1517 log.go:172] (0xc0005b1900) (5) Data frame handling\nI0425 00:23:24.575225 1517 log.go:172] (0xc0005b1900) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:23:24.626091 1517 log.go:172] (0xc000a7f1e0) Data frame received for 5\nI0425 00:23:24.626146 1517 log.go:172] (0xc0005b1900) (5) Data frame handling\nI0425 00:23:24.626209 1517 log.go:172] (0xc000a7f1e0) Data frame received for 3\nI0425 00:23:24.626253 1517 log.go:172] (0xc000405720) (3) Data frame handling\nI0425 00:23:24.626283 1517 log.go:172] (0xc000405720) (3) Data frame sent\nI0425 00:23:24.626307 1517 log.go:172] (0xc000a7f1e0) Data frame received for 3\nI0425 00:23:24.626328 1517 log.go:172] (0xc000405720) (3) Data frame handling\nI0425 00:23:24.627726 1517 log.go:172] (0xc000a7f1e0) Data frame received for 1\nI0425 00:23:24.627745 1517 log.go:172] (0xc0009e2500) (1) Data frame handling\nI0425 00:23:24.627763 1517 log.go:172] (0xc0009e2500) (1) Data frame sent\nI0425 00:23:24.627779 1517 log.go:172] (0xc000a7f1e0) (0xc0009e2500) Stream removed, broadcasting: 1\nI0425 00:23:24.628019 1517 log.go:172] (0xc000a7f1e0) Go away received\nI0425 00:23:24.628086 1517 log.go:172] (0xc000a7f1e0) (0xc0009e2500) Stream removed, broadcasting: 1\nI0425 00:23:24.628156 1517 log.go:172] (0xc000a7f1e0) (0xc000405720) Stream removed, broadcasting: 3\nI0425 00:23:24.628176 1517 log.go:172] (0xc000a7f1e0) (0xc0005b1900) Stream removed, broadcasting: 5\n" Apr 25 00:23:24.634: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:23:24.634: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:23:24.634: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:23:24.638: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 Apr 25 00:23:34.645: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:23:34.645: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:23:34.645: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:23:34.661: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:34.661: INFO: ss-0 latest-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC }] Apr 25 00:23:34.661: INFO: ss-1 latest-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:34.661: INFO: ss-2 latest-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:34.661: INFO: Apr 25 00:23:34.662: INFO: StatefulSet ss has not reached scale 0, at 3 Apr 25 00:23:35.681: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:35.681: INFO: ss-0 latest-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC }] Apr 25 00:23:35.682: INFO: ss-1 latest-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:35.682: INFO: ss-2 latest-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:35.682: INFO: Apr 25 00:23:35.682: INFO: StatefulSet ss has not reached scale 0, at 3 Apr 25 00:23:36.686: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:36.686: INFO: ss-0 latest-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:22:50 +0000 UTC }] Apr 25 00:23:36.686: INFO: ss-1 latest-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:36.686: INFO: ss-2 latest-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:36.686: INFO: Apr 25 00:23:36.686: INFO: StatefulSet ss has not reached scale 0, at 3 Apr 25 00:23:37.690: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:37.690: INFO: ss-1 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:37.690: INFO: ss-2 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:37.690: INFO: Apr 25 00:23:37.690: INFO: StatefulSet ss has not reached scale 0, at 2 Apr 25 00:23:38.695: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:38.695: INFO: ss-1 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:38.695: INFO: ss-2 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:38.695: INFO: Apr 25 00:23:38.695: INFO: StatefulSet ss has not reached scale 0, at 2 Apr 25 00:23:39.700: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:39.700: INFO: ss-1 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:39.700: INFO: ss-2 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:39.700: INFO: Apr 25 00:23:39.700: INFO: StatefulSet ss has not reached scale 0, at 2 Apr 25 00:23:40.705: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:40.705: INFO: ss-1 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:40.705: INFO: ss-2 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:40.705: INFO: Apr 25 00:23:40.705: INFO: StatefulSet ss has not reached scale 0, at 2 Apr 25 00:23:41.710: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:41.710: INFO: ss-1 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:41.710: INFO: ss-2 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:41.710: INFO: Apr 25 00:23:41.710: INFO: StatefulSet ss has not reached scale 0, at 2 Apr 25 00:23:42.730: INFO: POD NODE PHASE GRACE CONDITIONS Apr 25 00:23:42.730: INFO: ss-1 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:24 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:42.731: INFO: ss-2 latest-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-25 00:23:13 +0000 UTC }] Apr 25 00:23:42.731: INFO: Apr 25 00:23:42.731: INFO: StatefulSet ss has not reached scale 0, at 2 Apr 25 00:23:43.735: INFO: Verifying statefulset ss doesn't scale past 0 for another 921.102808ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-734 Apr 25 00:23:44.739: INFO: Scaling statefulset ss to 0 Apr 25 00:23:44.748: INFO: Waiting for statefulset status.replicas updated to 0 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:110 Apr 25 00:23:44.751: INFO: Deleting all statefulset in ns statefulset-734 Apr 25 00:23:44.754: INFO: Scaling statefulset ss to 0 Apr 25 00:23:44.762: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:23:44.764: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:23:44.775: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-734" for this suite. • [SLOW TEST:54.725 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","total":275,"completed":144,"skipped":2508,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:23:44.783: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 25 00:23:48.882: INFO: Expected: &{} to match Container's Termination Message: -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:23:48.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-6641" for this suite. •{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":275,"completed":145,"skipped":2576,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:23:48.924: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:23:49.018: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-1910d55b-f31a-4021-928f-d5324ea444be" in namespace "security-context-test-4855" to be "Succeeded or Failed" Apr 25 00:23:49.024: INFO: Pod "busybox-readonly-false-1910d55b-f31a-4021-928f-d5324ea444be": Phase="Pending", Reason="", readiness=false. Elapsed: 5.642089ms Apr 25 00:23:51.065: INFO: Pod "busybox-readonly-false-1910d55b-f31a-4021-928f-d5324ea444be": Phase="Pending", Reason="", readiness=false. Elapsed: 2.046682259s Apr 25 00:23:53.069: INFO: Pod "busybox-readonly-false-1910d55b-f31a-4021-928f-d5324ea444be": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.050801402s Apr 25 00:23:53.069: INFO: Pod "busybox-readonly-false-1910d55b-f31a-4021-928f-d5324ea444be" satisfied condition "Succeeded or Failed" [AfterEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:23:53.069: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-4855" for this suite. •{"msg":"PASSED [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","total":275,"completed":146,"skipped":2598,"failed":0} S ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:23:53.078: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod pod-subpath-test-secret-b8dm STEP: Creating a pod to test atomic-volume-subpath Apr 25 00:23:53.176: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-b8dm" in namespace "subpath-769" to be "Succeeded or Failed" Apr 25 00:23:53.180: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Pending", Reason="", readiness=false. Elapsed: 4.460819ms Apr 25 00:23:55.188: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01239277s Apr 25 00:23:57.192: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 4.01615476s Apr 25 00:23:59.196: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 6.020509715s Apr 25 00:24:01.200: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 8.024152355s Apr 25 00:24:03.204: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 10.028364727s Apr 25 00:24:05.212: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 12.036659857s Apr 25 00:24:07.216: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 14.040507124s Apr 25 00:24:09.221: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 16.045729247s Apr 25 00:24:11.225: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 18.049762966s Apr 25 00:24:13.230: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 20.053869812s Apr 25 00:24:15.251: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Running", Reason="", readiness=true. Elapsed: 22.075734767s Apr 25 00:24:17.257: INFO: Pod "pod-subpath-test-secret-b8dm": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.081709307s STEP: Saw pod success Apr 25 00:24:17.257: INFO: Pod "pod-subpath-test-secret-b8dm" satisfied condition "Succeeded or Failed" Apr 25 00:24:17.260: INFO: Trying to get logs from node latest-worker pod pod-subpath-test-secret-b8dm container test-container-subpath-secret-b8dm: STEP: delete the pod Apr 25 00:24:17.298: INFO: Waiting for pod pod-subpath-test-secret-b8dm to disappear Apr 25 00:24:17.302: INFO: Pod pod-subpath-test-secret-b8dm no longer exists STEP: Deleting pod pod-subpath-test-secret-b8dm Apr 25 00:24:17.302: INFO: Deleting pod "pod-subpath-test-secret-b8dm" in namespace "subpath-769" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:24:17.304: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-769" for this suite. • [SLOW TEST:24.232 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]","total":275,"completed":147,"skipped":2599,"failed":0} SSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:24:17.310: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:24:17.875: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:24:19.884: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371057, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371057, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371057, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371057, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:24:22.900: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should include webhook resources in discovery documents [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: fetching the /apis discovery document STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document STEP: fetching the /apis/admissionregistration.k8s.io discovery document STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:24:22.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-4666" for this suite. STEP: Destroying namespace "webhook-4666-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:5.699 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should include webhook resources in discovery documents [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":275,"completed":148,"skipped":2604,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:24:23.010: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 25 00:24:27.109: INFO: Expected: &{OK} to match Container's Termination Message: OK -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:24:27.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-5445" for this suite. •{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":275,"completed":149,"skipped":2621,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-scheduling] LimitRange /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:24:27.148: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename limitrange STEP: Waiting for a default service account to be provisioned in namespace [It] should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a LimitRange STEP: Setting up watch STEP: Submitting a LimitRange Apr 25 00:24:27.281: INFO: observed the limitRanges list STEP: Verifying LimitRange creation was observed STEP: Fetching the LimitRange to ensure it has proper values Apr 25 00:24:27.290: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] Apr 25 00:24:27.290: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Creating a Pod with no resource requirements STEP: Ensuring Pod has resource requirements applied from LimitRange Apr 25 00:24:27.330: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] Apr 25 00:24:27.330: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Creating a Pod with partial resource requirements STEP: Ensuring Pod has merged resource requirements applied from LimitRange Apr 25 00:24:27.381: INFO: Verifying requests: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] Apr 25 00:24:27.381: INFO: Verifying limits: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Failing to create a Pod with less than min resources STEP: Failing to create a Pod with more than max resources STEP: Updating a LimitRange STEP: Verifying LimitRange updating is effective STEP: Creating a Pod with less than former min resources STEP: Failing to create a Pod with more than max resources STEP: Deleting a LimitRange STEP: Verifying the LimitRange was deleted Apr 25 00:24:34.470: INFO: limitRange is already deleted STEP: Creating a Pod with more than former max resources [AfterEach] [sig-scheduling] LimitRange /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:24:34.477: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "limitrange-8301" for this suite. • [SLOW TEST:7.572 seconds] [sig-scheduling] LimitRange /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]","total":275,"completed":150,"skipped":2637,"failed":0} SSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:24:34.721: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: udp [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Performing setup for networking test in namespace pod-network-test-3169 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 25 00:24:35.045: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Apr 25 00:24:35.079: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:24:37.086: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:24:39.244: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:24:41.084: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:24:43.083: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:24:45.083: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:24:47.083: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:24:49.083: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:24:51.083: INFO: The status of Pod netserver-0 is Running (Ready = true) Apr 25 00:24:51.089: INFO: The status of Pod netserver-1 is Running (Ready = false) Apr 25 00:24:53.094: INFO: The status of Pod netserver-1 is Running (Ready = false) Apr 25 00:24:55.094: INFO: The status of Pod netserver-1 is Running (Ready = false) Apr 25 00:24:57.094: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Apr 25 00:25:01.115: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.212:8080/dial?request=hostname&protocol=udp&host=10.244.2.168&port=8081&tries=1'] Namespace:pod-network-test-3169 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:25:01.115: INFO: >>> kubeConfig: /root/.kube/config I0425 00:25:01.150649 8 log.go:172] (0xc0028de2c0) (0xc0017328c0) Create stream I0425 00:25:01.150675 8 log.go:172] (0xc0028de2c0) (0xc0017328c0) Stream added, broadcasting: 1 I0425 00:25:01.152693 8 log.go:172] (0xc0028de2c0) Reply frame received for 1 I0425 00:25:01.152739 8 log.go:172] (0xc0028de2c0) (0xc00233e280) Create stream I0425 00:25:01.152761 8 log.go:172] (0xc0028de2c0) (0xc00233e280) Stream added, broadcasting: 3 I0425 00:25:01.154241 8 log.go:172] (0xc0028de2c0) Reply frame received for 3 I0425 00:25:01.154281 8 log.go:172] (0xc0028de2c0) (0xc001732960) Create stream I0425 00:25:01.154297 8 log.go:172] (0xc0028de2c0) (0xc001732960) Stream added, broadcasting: 5 I0425 00:25:01.155499 8 log.go:172] (0xc0028de2c0) Reply frame received for 5 I0425 00:25:01.253768 8 log.go:172] (0xc0028de2c0) Data frame received for 3 I0425 00:25:01.253804 8 log.go:172] (0xc00233e280) (3) Data frame handling I0425 00:25:01.253828 8 log.go:172] (0xc00233e280) (3) Data frame sent I0425 00:25:01.254048 8 log.go:172] (0xc0028de2c0) Data frame received for 5 I0425 00:25:01.254061 8 log.go:172] (0xc001732960) (5) Data frame handling I0425 00:25:01.254085 8 log.go:172] (0xc0028de2c0) Data frame received for 3 I0425 00:25:01.254096 8 log.go:172] (0xc00233e280) (3) Data frame handling I0425 00:25:01.255564 8 log.go:172] (0xc0028de2c0) Data frame received for 1 I0425 00:25:01.255585 8 log.go:172] (0xc0017328c0) (1) Data frame handling I0425 00:25:01.255597 8 log.go:172] (0xc0017328c0) (1) Data frame sent I0425 00:25:01.255621 8 log.go:172] (0xc0028de2c0) (0xc0017328c0) Stream removed, broadcasting: 1 I0425 00:25:01.255746 8 log.go:172] (0xc0028de2c0) (0xc0017328c0) Stream removed, broadcasting: 1 I0425 00:25:01.255773 8 log.go:172] (0xc0028de2c0) (0xc00233e280) Stream removed, broadcasting: 3 I0425 00:25:01.255785 8 log.go:172] (0xc0028de2c0) (0xc001732960) Stream removed, broadcasting: 5 Apr 25 00:25:01.255: INFO: Waiting for responses: map[] I0425 00:25:01.256020 8 log.go:172] (0xc0028de2c0) Go away received Apr 25 00:25:01.260: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.212:8080/dial?request=hostname&protocol=udp&host=10.244.1.211&port=8081&tries=1'] Namespace:pod-network-test-3169 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:25:01.260: INFO: >>> kubeConfig: /root/.kube/config I0425 00:25:01.283801 8 log.go:172] (0xc0017ba4d0) (0xc001e60820) Create stream I0425 00:25:01.283826 8 log.go:172] (0xc0017ba4d0) (0xc001e60820) Stream added, broadcasting: 1 I0425 00:25:01.285581 8 log.go:172] (0xc0017ba4d0) Reply frame received for 1 I0425 00:25:01.285610 8 log.go:172] (0xc0017ba4d0) (0xc00233e500) Create stream I0425 00:25:01.285629 8 log.go:172] (0xc0017ba4d0) (0xc00233e500) Stream added, broadcasting: 3 I0425 00:25:01.286562 8 log.go:172] (0xc0017ba4d0) Reply frame received for 3 I0425 00:25:01.286605 8 log.go:172] (0xc0017ba4d0) (0xc001732b40) Create stream I0425 00:25:01.286616 8 log.go:172] (0xc0017ba4d0) (0xc001732b40) Stream added, broadcasting: 5 I0425 00:25:01.287351 8 log.go:172] (0xc0017ba4d0) Reply frame received for 5 I0425 00:25:01.341014 8 log.go:172] (0xc0017ba4d0) Data frame received for 3 I0425 00:25:01.341063 8 log.go:172] (0xc00233e500) (3) Data frame handling I0425 00:25:01.341103 8 log.go:172] (0xc00233e500) (3) Data frame sent I0425 00:25:01.341642 8 log.go:172] (0xc0017ba4d0) Data frame received for 3 I0425 00:25:01.341674 8 log.go:172] (0xc00233e500) (3) Data frame handling I0425 00:25:01.341716 8 log.go:172] (0xc0017ba4d0) Data frame received for 5 I0425 00:25:01.341794 8 log.go:172] (0xc001732b40) (5) Data frame handling I0425 00:25:01.343581 8 log.go:172] (0xc0017ba4d0) Data frame received for 1 I0425 00:25:01.343617 8 log.go:172] (0xc001e60820) (1) Data frame handling I0425 00:25:01.343656 8 log.go:172] (0xc001e60820) (1) Data frame sent I0425 00:25:01.343686 8 log.go:172] (0xc0017ba4d0) (0xc001e60820) Stream removed, broadcasting: 1 I0425 00:25:01.343713 8 log.go:172] (0xc0017ba4d0) Go away received I0425 00:25:01.343936 8 log.go:172] (0xc0017ba4d0) (0xc001e60820) Stream removed, broadcasting: 1 I0425 00:25:01.343969 8 log.go:172] (0xc0017ba4d0) (0xc00233e500) Stream removed, broadcasting: 3 I0425 00:25:01.343991 8 log.go:172] (0xc0017ba4d0) (0xc001732b40) Stream removed, broadcasting: 5 Apr 25 00:25:01.344: INFO: Waiting for responses: map[] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:01.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-3169" for this suite. • [SLOW TEST:26.632 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for intra-pod communication: udp [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]","total":275,"completed":151,"skipped":2641,"failed":0} SSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:01.354: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should be able to change the type from ExternalName to NodePort [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a service externalname-service with the type=ExternalName in namespace services-8711 STEP: changing the ExternalName service to type=NodePort STEP: creating replication controller externalname-service in namespace services-8711 I0425 00:25:01.478171 8 runners.go:190] Created replication controller with name: externalname-service, namespace: services-8711, replica count: 2 I0425 00:25:04.528648 8 runners.go:190] externalname-service Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0425 00:25:07.528882 8 runners.go:190] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Apr 25 00:25:07.528: INFO: Creating new exec pod Apr 25 00:25:12.563: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-8711 execpod9mnhs -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80' Apr 25 00:25:12.786: INFO: stderr: "I0425 00:25:12.685726 1540 log.go:172] (0xc000a020b0) (0xc000a3c000) Create stream\nI0425 00:25:12.685935 1540 log.go:172] (0xc000a020b0) (0xc000a3c000) Stream added, broadcasting: 1\nI0425 00:25:12.688323 1540 log.go:172] (0xc000a020b0) Reply frame received for 1\nI0425 00:25:12.688363 1540 log.go:172] (0xc000a020b0) (0xc00097a000) Create stream\nI0425 00:25:12.688374 1540 log.go:172] (0xc000a020b0) (0xc00097a000) Stream added, broadcasting: 3\nI0425 00:25:12.689825 1540 log.go:172] (0xc000a020b0) Reply frame received for 3\nI0425 00:25:12.689874 1540 log.go:172] (0xc000a020b0) (0xc000699360) Create stream\nI0425 00:25:12.689891 1540 log.go:172] (0xc000a020b0) (0xc000699360) Stream added, broadcasting: 5\nI0425 00:25:12.690946 1540 log.go:172] (0xc000a020b0) Reply frame received for 5\nI0425 00:25:12.778868 1540 log.go:172] (0xc000a020b0) Data frame received for 3\nI0425 00:25:12.778908 1540 log.go:172] (0xc00097a000) (3) Data frame handling\nI0425 00:25:12.778932 1540 log.go:172] (0xc000a020b0) Data frame received for 5\nI0425 00:25:12.778941 1540 log.go:172] (0xc000699360) (5) Data frame handling\nI0425 00:25:12.778953 1540 log.go:172] (0xc000699360) (5) Data frame sent\nI0425 00:25:12.778967 1540 log.go:172] (0xc000a020b0) Data frame received for 5\n+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\nI0425 00:25:12.778988 1540 log.go:172] (0xc000699360) (5) Data frame handling\nI0425 00:25:12.780986 1540 log.go:172] (0xc000a020b0) Data frame received for 1\nI0425 00:25:12.781034 1540 log.go:172] (0xc000a3c000) (1) Data frame handling\nI0425 00:25:12.781071 1540 log.go:172] (0xc000a3c000) (1) Data frame sent\nI0425 00:25:12.781275 1540 log.go:172] (0xc000a020b0) (0xc000a3c000) Stream removed, broadcasting: 1\nI0425 00:25:12.781357 1540 log.go:172] (0xc000a020b0) Go away received\nI0425 00:25:12.781868 1540 log.go:172] (0xc000a020b0) (0xc000a3c000) Stream removed, broadcasting: 1\nI0425 00:25:12.781900 1540 log.go:172] (0xc000a020b0) (0xc00097a000) Stream removed, broadcasting: 3\nI0425 00:25:12.781914 1540 log.go:172] (0xc000a020b0) (0xc000699360) Stream removed, broadcasting: 5\n" Apr 25 00:25:12.786: INFO: stdout: "" Apr 25 00:25:12.786: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-8711 execpod9mnhs -- /bin/sh -x -c nc -zv -t -w 2 10.96.125.188 80' Apr 25 00:25:12.991: INFO: stderr: "I0425 00:25:12.911393 1561 log.go:172] (0xc0000e84d0) (0xc0009c6000) Create stream\nI0425 00:25:12.911472 1561 log.go:172] (0xc0000e84d0) (0xc0009c6000) Stream added, broadcasting: 1\nI0425 00:25:12.915119 1561 log.go:172] (0xc0000e84d0) Reply frame received for 1\nI0425 00:25:12.915171 1561 log.go:172] (0xc0000e84d0) (0xc00090c000) Create stream\nI0425 00:25:12.915192 1561 log.go:172] (0xc0000e84d0) (0xc00090c000) Stream added, broadcasting: 3\nI0425 00:25:12.916388 1561 log.go:172] (0xc0000e84d0) Reply frame received for 3\nI0425 00:25:12.916427 1561 log.go:172] (0xc0000e84d0) (0xc0009c60a0) Create stream\nI0425 00:25:12.916439 1561 log.go:172] (0xc0000e84d0) (0xc0009c60a0) Stream added, broadcasting: 5\nI0425 00:25:12.917812 1561 log.go:172] (0xc0000e84d0) Reply frame received for 5\nI0425 00:25:12.984455 1561 log.go:172] (0xc0000e84d0) Data frame received for 3\nI0425 00:25:12.984503 1561 log.go:172] (0xc00090c000) (3) Data frame handling\nI0425 00:25:12.984531 1561 log.go:172] (0xc0000e84d0) Data frame received for 5\nI0425 00:25:12.984543 1561 log.go:172] (0xc0009c60a0) (5) Data frame handling\nI0425 00:25:12.984562 1561 log.go:172] (0xc0009c60a0) (5) Data frame sent\nI0425 00:25:12.984573 1561 log.go:172] (0xc0000e84d0) Data frame received for 5\nI0425 00:25:12.984583 1561 log.go:172] (0xc0009c60a0) (5) Data frame handling\n+ nc -zv -t -w 2 10.96.125.188 80\nConnection to 10.96.125.188 80 port [tcp/http] succeeded!\nI0425 00:25:12.985958 1561 log.go:172] (0xc0000e84d0) Data frame received for 1\nI0425 00:25:12.985980 1561 log.go:172] (0xc0009c6000) (1) Data frame handling\nI0425 00:25:12.985996 1561 log.go:172] (0xc0009c6000) (1) Data frame sent\nI0425 00:25:12.986021 1561 log.go:172] (0xc0000e84d0) (0xc0009c6000) Stream removed, broadcasting: 1\nI0425 00:25:12.986048 1561 log.go:172] (0xc0000e84d0) Go away received\nI0425 00:25:12.986379 1561 log.go:172] (0xc0000e84d0) (0xc0009c6000) Stream removed, broadcasting: 1\nI0425 00:25:12.986402 1561 log.go:172] (0xc0000e84d0) (0xc00090c000) Stream removed, broadcasting: 3\nI0425 00:25:12.986421 1561 log.go:172] (0xc0000e84d0) (0xc0009c60a0) Stream removed, broadcasting: 5\n" Apr 25 00:25:12.991: INFO: stdout: "" Apr 25 00:25:12.991: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-8711 execpod9mnhs -- /bin/sh -x -c nc -zv -t -w 2 172.17.0.13 32649' Apr 25 00:25:13.173: INFO: stderr: "I0425 00:25:13.112009 1580 log.go:172] (0xc0009dc6e0) (0xc0008cc000) Create stream\nI0425 00:25:13.112074 1580 log.go:172] (0xc0009dc6e0) (0xc0008cc000) Stream added, broadcasting: 1\nI0425 00:25:13.114800 1580 log.go:172] (0xc0009dc6e0) Reply frame received for 1\nI0425 00:25:13.114861 1580 log.go:172] (0xc0009dc6e0) (0xc0008bc000) Create stream\nI0425 00:25:13.114893 1580 log.go:172] (0xc0009dc6e0) (0xc0008bc000) Stream added, broadcasting: 3\nI0425 00:25:13.115913 1580 log.go:172] (0xc0009dc6e0) Reply frame received for 3\nI0425 00:25:13.115975 1580 log.go:172] (0xc0009dc6e0) (0xc0008bc0a0) Create stream\nI0425 00:25:13.115999 1580 log.go:172] (0xc0009dc6e0) (0xc0008bc0a0) Stream added, broadcasting: 5\nI0425 00:25:13.116812 1580 log.go:172] (0xc0009dc6e0) Reply frame received for 5\nI0425 00:25:13.166092 1580 log.go:172] (0xc0009dc6e0) Data frame received for 5\nI0425 00:25:13.166152 1580 log.go:172] (0xc0008bc0a0) (5) Data frame handling\nI0425 00:25:13.166190 1580 log.go:172] (0xc0008bc0a0) (5) Data frame sent\nI0425 00:25:13.166211 1580 log.go:172] (0xc0009dc6e0) Data frame received for 5\nI0425 00:25:13.166226 1580 log.go:172] (0xc0008bc0a0) (5) Data frame handling\n+ nc -zv -t -w 2 172.17.0.13 32649\nConnection to 172.17.0.13 32649 port [tcp/32649] succeeded!\nI0425 00:25:13.166286 1580 log.go:172] (0xc0008bc0a0) (5) Data frame sent\nI0425 00:25:13.166493 1580 log.go:172] (0xc0009dc6e0) Data frame received for 5\nI0425 00:25:13.166525 1580 log.go:172] (0xc0009dc6e0) Data frame received for 3\nI0425 00:25:13.166551 1580 log.go:172] (0xc0008bc000) (3) Data frame handling\nI0425 00:25:13.166574 1580 log.go:172] (0xc0008bc0a0) (5) Data frame handling\nI0425 00:25:13.168212 1580 log.go:172] (0xc0009dc6e0) Data frame received for 1\nI0425 00:25:13.168230 1580 log.go:172] (0xc0008cc000) (1) Data frame handling\nI0425 00:25:13.168240 1580 log.go:172] (0xc0008cc000) (1) Data frame sent\nI0425 00:25:13.168260 1580 log.go:172] (0xc0009dc6e0) (0xc0008cc000) Stream removed, broadcasting: 1\nI0425 00:25:13.168455 1580 log.go:172] (0xc0009dc6e0) Go away received\nI0425 00:25:13.168706 1580 log.go:172] (0xc0009dc6e0) (0xc0008cc000) Stream removed, broadcasting: 1\nI0425 00:25:13.168733 1580 log.go:172] (0xc0009dc6e0) (0xc0008bc000) Stream removed, broadcasting: 3\nI0425 00:25:13.168748 1580 log.go:172] (0xc0009dc6e0) (0xc0008bc0a0) Stream removed, broadcasting: 5\n" Apr 25 00:25:13.173: INFO: stdout: "" Apr 25 00:25:13.173: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-8711 execpod9mnhs -- /bin/sh -x -c nc -zv -t -w 2 172.17.0.12 32649' Apr 25 00:25:13.389: INFO: stderr: "I0425 00:25:13.300721 1603 log.go:172] (0xc000b9d290) (0xc000994820) Create stream\nI0425 00:25:13.300801 1603 log.go:172] (0xc000b9d290) (0xc000994820) Stream added, broadcasting: 1\nI0425 00:25:13.306797 1603 log.go:172] (0xc000b9d290) Reply frame received for 1\nI0425 00:25:13.306855 1603 log.go:172] (0xc000b9d290) (0xc00079b540) Create stream\nI0425 00:25:13.306878 1603 log.go:172] (0xc000b9d290) (0xc00079b540) Stream added, broadcasting: 3\nI0425 00:25:13.308012 1603 log.go:172] (0xc000b9d290) Reply frame received for 3\nI0425 00:25:13.308052 1603 log.go:172] (0xc000b9d290) (0xc00059a960) Create stream\nI0425 00:25:13.308066 1603 log.go:172] (0xc000b9d290) (0xc00059a960) Stream added, broadcasting: 5\nI0425 00:25:13.309019 1603 log.go:172] (0xc000b9d290) Reply frame received for 5\nI0425 00:25:13.381770 1603 log.go:172] (0xc000b9d290) Data frame received for 5\nI0425 00:25:13.381806 1603 log.go:172] (0xc00059a960) (5) Data frame handling\nI0425 00:25:13.381845 1603 log.go:172] (0xc00059a960) (5) Data frame sent\nI0425 00:25:13.381859 1603 log.go:172] (0xc000b9d290) Data frame received for 5\nI0425 00:25:13.381870 1603 log.go:172] (0xc00059a960) (5) Data frame handling\n+ nc -zv -t -w 2 172.17.0.12 32649\nConnection to 172.17.0.12 32649 port [tcp/32649] succeeded!\nI0425 00:25:13.381887 1603 log.go:172] (0xc000b9d290) Data frame received for 3\nI0425 00:25:13.381923 1603 log.go:172] (0xc00079b540) (3) Data frame handling\nI0425 00:25:13.383475 1603 log.go:172] (0xc000b9d290) Data frame received for 1\nI0425 00:25:13.383503 1603 log.go:172] (0xc000994820) (1) Data frame handling\nI0425 00:25:13.383551 1603 log.go:172] (0xc000994820) (1) Data frame sent\nI0425 00:25:13.383587 1603 log.go:172] (0xc000b9d290) (0xc000994820) Stream removed, broadcasting: 1\nI0425 00:25:13.383615 1603 log.go:172] (0xc000b9d290) Go away received\nI0425 00:25:13.384131 1603 log.go:172] (0xc000b9d290) (0xc000994820) Stream removed, broadcasting: 1\nI0425 00:25:13.384156 1603 log.go:172] (0xc000b9d290) (0xc00079b540) Stream removed, broadcasting: 3\nI0425 00:25:13.384167 1603 log.go:172] (0xc000b9d290) (0xc00059a960) Stream removed, broadcasting: 5\n" Apr 25 00:25:13.389: INFO: stdout: "" Apr 25 00:25:13.389: INFO: Cleaning up the ExternalName to NodePort test service [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:13.437: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8711" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 • [SLOW TEST:12.098 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from ExternalName to NodePort [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","total":275,"completed":152,"skipped":2654,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:13.452: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating the pod Apr 25 00:25:18.043: INFO: Successfully updated pod "annotationupdateafeacc44-c230-480c-b5d1-d07b7f589afb" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:20.064: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5312" for this suite. • [SLOW TEST:6.621 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","total":275,"completed":153,"skipped":2672,"failed":0} SSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:20.074: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [BeforeEach] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:271 [It] should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a replication controller Apr 25 00:25:20.102: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8516' Apr 25 00:25:20.457: INFO: stderr: "" Apr 25 00:25:20.457: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 25 00:25:20.457: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8516' Apr 25 00:25:20.580: INFO: stderr: "" Apr 25 00:25:20.580: INFO: stdout: "update-demo-nautilus-bxfhz update-demo-nautilus-jxwx4 " Apr 25 00:25:20.580: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-bxfhz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:20.672: INFO: stderr: "" Apr 25 00:25:20.672: INFO: stdout: "" Apr 25 00:25:20.672: INFO: update-demo-nautilus-bxfhz is created but not running Apr 25 00:25:25.672: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8516' Apr 25 00:25:25.779: INFO: stderr: "" Apr 25 00:25:25.779: INFO: stdout: "update-demo-nautilus-bxfhz update-demo-nautilus-jxwx4 " Apr 25 00:25:25.779: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-bxfhz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:25.872: INFO: stderr: "" Apr 25 00:25:25.872: INFO: stdout: "true" Apr 25 00:25:25.872: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-bxfhz -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:25.964: INFO: stderr: "" Apr 25 00:25:25.964: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 25 00:25:25.964: INFO: validating pod update-demo-nautilus-bxfhz Apr 25 00:25:25.968: INFO: got data: { "image": "nautilus.jpg" } Apr 25 00:25:25.968: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 25 00:25:25.968: INFO: update-demo-nautilus-bxfhz is verified up and running Apr 25 00:25:25.968: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jxwx4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:26.063: INFO: stderr: "" Apr 25 00:25:26.063: INFO: stdout: "true" Apr 25 00:25:26.063: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jxwx4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:26.157: INFO: stderr: "" Apr 25 00:25:26.157: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 25 00:25:26.157: INFO: validating pod update-demo-nautilus-jxwx4 Apr 25 00:25:26.161: INFO: got data: { "image": "nautilus.jpg" } Apr 25 00:25:26.161: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 25 00:25:26.161: INFO: update-demo-nautilus-jxwx4 is verified up and running STEP: scaling down the replication controller Apr 25 00:25:26.163: INFO: scanned /root for discovery docs: Apr 25 00:25:26.163: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-8516' Apr 25 00:25:27.301: INFO: stderr: "" Apr 25 00:25:27.301: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 25 00:25:27.302: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8516' Apr 25 00:25:27.409: INFO: stderr: "" Apr 25 00:25:27.409: INFO: stdout: "update-demo-nautilus-bxfhz update-demo-nautilus-jxwx4 " STEP: Replicas for name=update-demo: expected=1 actual=2 Apr 25 00:25:32.409: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8516' Apr 25 00:25:32.504: INFO: stderr: "" Apr 25 00:25:32.504: INFO: stdout: "update-demo-nautilus-bxfhz update-demo-nautilus-jxwx4 " STEP: Replicas for name=update-demo: expected=1 actual=2 Apr 25 00:25:37.505: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8516' Apr 25 00:25:37.601: INFO: stderr: "" Apr 25 00:25:37.601: INFO: stdout: "update-demo-nautilus-jxwx4 " Apr 25 00:25:37.601: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jxwx4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:37.696: INFO: stderr: "" Apr 25 00:25:37.696: INFO: stdout: "true" Apr 25 00:25:37.696: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jxwx4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:37.788: INFO: stderr: "" Apr 25 00:25:37.788: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 25 00:25:37.788: INFO: validating pod update-demo-nautilus-jxwx4 Apr 25 00:25:37.792: INFO: got data: { "image": "nautilus.jpg" } Apr 25 00:25:37.792: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 25 00:25:37.792: INFO: update-demo-nautilus-jxwx4 is verified up and running STEP: scaling up the replication controller Apr 25 00:25:37.794: INFO: scanned /root for discovery docs: Apr 25 00:25:37.794: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-8516' Apr 25 00:25:38.920: INFO: stderr: "" Apr 25 00:25:38.920: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 25 00:25:38.920: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8516' Apr 25 00:25:39.016: INFO: stderr: "" Apr 25 00:25:39.016: INFO: stdout: "update-demo-nautilus-drhn5 update-demo-nautilus-jxwx4 " Apr 25 00:25:39.016: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-drhn5 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:39.104: INFO: stderr: "" Apr 25 00:25:39.104: INFO: stdout: "" Apr 25 00:25:39.104: INFO: update-demo-nautilus-drhn5 is created but not running Apr 25 00:25:44.104: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8516' Apr 25 00:25:44.201: INFO: stderr: "" Apr 25 00:25:44.201: INFO: stdout: "update-demo-nautilus-drhn5 update-demo-nautilus-jxwx4 " Apr 25 00:25:44.201: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-drhn5 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:44.299: INFO: stderr: "" Apr 25 00:25:44.299: INFO: stdout: "true" Apr 25 00:25:44.299: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-drhn5 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:44.397: INFO: stderr: "" Apr 25 00:25:44.397: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 25 00:25:44.397: INFO: validating pod update-demo-nautilus-drhn5 Apr 25 00:25:44.402: INFO: got data: { "image": "nautilus.jpg" } Apr 25 00:25:44.402: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 25 00:25:44.402: INFO: update-demo-nautilus-drhn5 is verified up and running Apr 25 00:25:44.402: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jxwx4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:44.482: INFO: stderr: "" Apr 25 00:25:44.482: INFO: stdout: "true" Apr 25 00:25:44.482: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jxwx4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8516' Apr 25 00:25:44.575: INFO: stderr: "" Apr 25 00:25:44.575: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 25 00:25:44.575: INFO: validating pod update-demo-nautilus-jxwx4 Apr 25 00:25:44.579: INFO: got data: { "image": "nautilus.jpg" } Apr 25 00:25:44.579: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 25 00:25:44.579: INFO: update-demo-nautilus-jxwx4 is verified up and running STEP: using delete to clean up resources Apr 25 00:25:44.579: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8516' Apr 25 00:25:44.676: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 25 00:25:44.676: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Apr 25 00:25:44.676: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-8516' Apr 25 00:25:44.816: INFO: stderr: "No resources found in kubectl-8516 namespace.\n" Apr 25 00:25:44.816: INFO: stdout: "" Apr 25 00:25:44.816: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-8516 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 25 00:25:44.904: INFO: stderr: "" Apr 25 00:25:44.904: INFO: stdout: "update-demo-nautilus-drhn5\nupdate-demo-nautilus-jxwx4\n" Apr 25 00:25:45.404: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-8516' Apr 25 00:25:45.505: INFO: stderr: "No resources found in kubectl-8516 namespace.\n" Apr 25 00:25:45.505: INFO: stdout: "" Apr 25 00:25:45.505: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-8516 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 25 00:25:45.595: INFO: stderr: "" Apr 25 00:25:45.595: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:45.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8516" for this suite. • [SLOW TEST:25.528 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:269 should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance]","total":275,"completed":154,"skipped":2683,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:45.602: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:25:45.686: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91" in namespace "projected-2704" to be "Succeeded or Failed" Apr 25 00:25:45.731: INFO: Pod "downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91": Phase="Pending", Reason="", readiness=false. Elapsed: 44.762726ms Apr 25 00:25:47.735: INFO: Pod "downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04896076s Apr 25 00:25:49.740: INFO: Pod "downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.053948733s STEP: Saw pod success Apr 25 00:25:49.740: INFO: Pod "downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91" satisfied condition "Succeeded or Failed" Apr 25 00:25:49.743: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91 container client-container: STEP: delete the pod Apr 25 00:25:49.776: INFO: Waiting for pod downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91 to disappear Apr 25 00:25:49.787: INFO: Pod downwardapi-volume-4e7613aa-df30-48ec-b940-fd8797921c91 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:49.787: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2704" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":275,"completed":155,"skipped":2694,"failed":0} ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:49.795: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] getting/updating/patching custom resource definition status sub-resource works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:25:49.834: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:50.406: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-2736" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance]","total":275,"completed":156,"skipped":2694,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-network] Services should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:50.424: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:50.558: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3006" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 •{"msg":"PASSED [sig-network] Services should provide secure master service [Conformance]","total":275,"completed":157,"skipped":2706,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:50.567: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the container STEP: wait for the container to reach Failed STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 25 00:25:54.678: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:54.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-4923" for this suite. •{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":275,"completed":158,"skipped":2757,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:54.850: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:25:54.961: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac" in namespace "projected-867" to be "Succeeded or Failed" Apr 25 00:25:54.996: INFO: Pod "downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac": Phase="Pending", Reason="", readiness=false. Elapsed: 34.504105ms Apr 25 00:25:57.000: INFO: Pod "downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.038784617s Apr 25 00:25:59.004: INFO: Pod "downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.043079721s STEP: Saw pod success Apr 25 00:25:59.005: INFO: Pod "downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac" satisfied condition "Succeeded or Failed" Apr 25 00:25:59.009: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac container client-container: STEP: delete the pod Apr 25 00:25:59.025: INFO: Waiting for pod downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac to disappear Apr 25 00:25:59.029: INFO: Pod downwardapi-volume-ebd801be-ce40-4014-ae0d-4fd2a66625ac no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:25:59.029: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-867" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":275,"completed":159,"skipped":2969,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:25:59.036: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-upd-8432562e-d417-4216-8e92-13506cc712d7 STEP: Creating the pod STEP: Updating configmap configmap-test-upd-8432562e-d417-4216-8e92-13506cc712d7 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:26:07.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4682" for this suite. • [SLOW TEST:8.293 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":275,"completed":160,"skipped":3010,"failed":0} [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:26:07.328: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods Set QOS Class /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:157 [It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying QOS class is set on the pod [AfterEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:26:07.424: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8004" for this suite. •{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","total":275,"completed":161,"skipped":3010,"failed":0} SSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:26:07.435: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: retrieving the pod Apr 25 00:26:11.553: INFO: &Pod{ObjectMeta:{send-events-e358d73f-9ddf-43ee-a153-29a75f0b8a4c events-1622 /api/v1/namespaces/events-1622/pods/send-events-e358d73f-9ddf-43ee-a153-29a75f0b8a4c 9b90c476-5af9-4b4e-ae6e-29627faf0fe9 10793901 0 2020-04-25 00:26:07 +0000 UTC map[name:foo time:509463545] map[] [] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-6mnj4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-6mnj4,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:p,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,Command:[],Args:[serve-hostname],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:,HostPort:0,ContainerPort:80,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-6mnj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:26:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:26:11 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:26:11 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:26:07 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.176,StartTime:2020-04-25 00:26:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:p,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:26:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,ImageID:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost@sha256:1d7f0d77a6f07fd507f147a38d06a7c8269ebabd4f923bfe46d4fb8b396a520c,ContainerID:containerd://5445a588705bf0f5a5d66c4e55f58589af149069290f22ff5bfb6dca6ccdf95e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.176,},},EphemeralContainerStatuses:[]ContainerStatus{},},} STEP: checking for scheduler event about the pod Apr 25 00:26:13.557: INFO: Saw scheduler event for our pod. STEP: checking for kubelet event about the pod Apr 25 00:26:15.562: INFO: Saw kubelet event for our pod. STEP: deleting the pod [AfterEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:26:15.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-1622" for this suite. • [SLOW TEST:8.169 seconds] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance]","total":275,"completed":162,"skipped":3024,"failed":0} SSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:26:15.605: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Apr 25 00:26:23.708: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 25 00:26:23.713: INFO: Pod pod-with-prestop-http-hook still exists Apr 25 00:26:25.713: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 25 00:26:25.716: INFO: Pod pod-with-prestop-http-hook still exists Apr 25 00:26:27.713: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 25 00:26:27.716: INFO: Pod pod-with-prestop-http-hook still exists Apr 25 00:26:29.713: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 25 00:26:29.718: INFO: Pod pod-with-prestop-http-hook still exists Apr 25 00:26:31.713: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 25 00:26:31.719: INFO: Pod pod-with-prestop-http-hook still exists Apr 25 00:26:33.714: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 25 00:26:33.718: INFO: Pod pod-with-prestop-http-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:26:33.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-9984" for this suite. • [SLOW TEST:18.128 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":275,"completed":163,"skipped":3033,"failed":0} [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:26:33.733: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward api env vars Apr 25 00:26:33.812: INFO: Waiting up to 5m0s for pod "downward-api-e574caf4-482a-45b5-b028-c399fd724021" in namespace "downward-api-5132" to be "Succeeded or Failed" Apr 25 00:26:33.825: INFO: Pod "downward-api-e574caf4-482a-45b5-b028-c399fd724021": Phase="Pending", Reason="", readiness=false. Elapsed: 13.23218ms Apr 25 00:26:35.829: INFO: Pod "downward-api-e574caf4-482a-45b5-b028-c399fd724021": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016962236s Apr 25 00:26:37.832: INFO: Pod "downward-api-e574caf4-482a-45b5-b028-c399fd724021": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020460209s STEP: Saw pod success Apr 25 00:26:37.832: INFO: Pod "downward-api-e574caf4-482a-45b5-b028-c399fd724021" satisfied condition "Succeeded or Failed" Apr 25 00:26:37.835: INFO: Trying to get logs from node latest-worker2 pod downward-api-e574caf4-482a-45b5-b028-c399fd724021 container dapi-container: STEP: delete the pod Apr 25 00:26:37.882: INFO: Waiting for pod downward-api-e574caf4-482a-45b5-b028-c399fd724021 to disappear Apr 25 00:26:37.896: INFO: Pod downward-api-e574caf4-482a-45b5-b028-c399fd724021 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:26:37.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5132" for this suite. •{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":275,"completed":164,"skipped":3033,"failed":0} SSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:26:37.919: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod pod-subpath-test-projected-slpj STEP: Creating a pod to test atomic-volume-subpath Apr 25 00:26:37.995: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-slpj" in namespace "subpath-9818" to be "Succeeded or Failed" Apr 25 00:26:37.998: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Pending", Reason="", readiness=false. Elapsed: 3.599984ms Apr 25 00:26:40.002: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007008951s Apr 25 00:26:42.005: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 4.010564509s Apr 25 00:26:44.018: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 6.023646733s Apr 25 00:26:46.022: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 8.027796886s Apr 25 00:26:48.027: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 10.032140759s Apr 25 00:26:50.031: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 12.036317189s Apr 25 00:26:52.035: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 14.040764592s Apr 25 00:26:54.040: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 16.045204064s Apr 25 00:26:56.044: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 18.049188217s Apr 25 00:26:58.048: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 20.053271656s Apr 25 00:27:00.052: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Running", Reason="", readiness=true. Elapsed: 22.057535121s Apr 25 00:27:02.057: INFO: Pod "pod-subpath-test-projected-slpj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.062286685s STEP: Saw pod success Apr 25 00:27:02.057: INFO: Pod "pod-subpath-test-projected-slpj" satisfied condition "Succeeded or Failed" Apr 25 00:27:02.060: INFO: Trying to get logs from node latest-worker2 pod pod-subpath-test-projected-slpj container test-container-subpath-projected-slpj: STEP: delete the pod Apr 25 00:27:02.097: INFO: Waiting for pod pod-subpath-test-projected-slpj to disappear Apr 25 00:27:02.128: INFO: Pod pod-subpath-test-projected-slpj no longer exists STEP: Deleting pod pod-subpath-test-projected-slpj Apr 25 00:27:02.128: INFO: Deleting pod "pod-subpath-test-projected-slpj" in namespace "subpath-9818" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:27:02.131: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-9818" for this suite. • [SLOW TEST:24.219 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":275,"completed":165,"skipped":3040,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:27:02.139: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:27:02.195: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0" in namespace "downward-api-6301" to be "Succeeded or Failed" Apr 25 00:27:02.248: INFO: Pod "downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0": Phase="Pending", Reason="", readiness=false. Elapsed: 52.433206ms Apr 25 00:27:04.252: INFO: Pod "downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.05673118s Apr 25 00:27:06.257: INFO: Pod "downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.061379099s STEP: Saw pod success Apr 25 00:27:06.257: INFO: Pod "downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0" satisfied condition "Succeeded or Failed" Apr 25 00:27:06.260: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0 container client-container: STEP: delete the pod Apr 25 00:27:06.276: INFO: Waiting for pod downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0 to disappear Apr 25 00:27:06.280: INFO: Pod downwardapi-volume-3de9d015-2122-41cd-843f-f399f19b8ba0 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:27:06.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6301" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","total":275,"completed":166,"skipped":3107,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:27:06.288: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a watch on configmaps with label A STEP: creating a watch on configmaps with label B STEP: creating a watch on configmaps with label A or B STEP: creating a configmap with label A and ensuring the correct watchers observe the notification Apr 25 00:27:06.340: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794207 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:27:06.340: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794207 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying configmap A and ensuring the correct watchers observe the notification Apr 25 00:27:16.348: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794257 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:27:16.348: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794257 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying configmap A again and ensuring the correct watchers observe the notification Apr 25 00:27:26.356: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794287 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:27:26.357: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794287 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: deleting configmap A and ensuring the correct watchers observe the notification Apr 25 00:27:36.363: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794317 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:27:36.363: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-a 660f74be-db45-491e-b819-6713e82696af 10794317 0 2020-04-25 00:27:06 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: creating a configmap with label B and ensuring the correct watchers observe the notification Apr 25 00:27:46.370: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-b 97a03130-8180-40b9-b662-f5c318c09d5f 10794347 0 2020-04-25 00:27:46 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:27:46.371: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-b 97a03130-8180-40b9-b662-f5c318c09d5f 10794347 0 2020-04-25 00:27:46 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} STEP: deleting configmap B and ensuring the correct watchers observe the notification Apr 25 00:27:56.378: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-b 97a03130-8180-40b9-b662-f5c318c09d5f 10794377 0 2020-04-25 00:27:46 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:27:56.378: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-2981 /api/v1/namespaces/watch-2981/configmaps/e2e-watch-test-configmap-b 97a03130-8180-40b9-b662-f5c318c09d5f 10794377 0 2020-04-25 00:27:46 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:28:06.378: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-2981" for this suite. • [SLOW TEST:60.148 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":275,"completed":167,"skipped":3139,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:28:06.437: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [BeforeEach] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1454 [It] should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: running the image docker.io/library/httpd:2.4.38-alpine Apr 25 00:28:06.483: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config run e2e-test-httpd-pod --image=docker.io/library/httpd:2.4.38-alpine --labels=run=e2e-test-httpd-pod --namespace=kubectl-1276' Apr 25 00:28:06.603: INFO: stderr: "" Apr 25 00:28:06.603: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: verifying the pod e2e-test-httpd-pod is running STEP: verifying the pod e2e-test-httpd-pod was created Apr 25 00:28:11.654: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pod e2e-test-httpd-pod --namespace=kubectl-1276 -o json' Apr 25 00:28:11.754: INFO: stderr: "" Apr 25 00:28:11.754: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2020-04-25T00:28:06Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-1276\",\n \"resourceVersion\": \"10794429\",\n \"selfLink\": \"/api/v1/namespaces/kubectl-1276/pods/e2e-test-httpd-pod\",\n \"uid\": \"4f16d299-f2d1-4ee8-824f-d14de6b7e4f6\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-xqc6j\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"latest-worker\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-xqc6j\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-xqc6j\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-25T00:28:06Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-25T00:28:09Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-25T00:28:09Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-25T00:28:06Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://1a07eb0a412549ac4821f5d0ae24d64945e4abf74cc73bb3a45ae1e07eb68fe5\",\n \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n \"imageID\": \"docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"started\": true,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2020-04-25T00:28:08Z\"\n }\n }\n }\n ],\n \"hostIP\": \"172.17.0.13\",\n \"phase\": \"Running\",\n \"podIP\": \"10.244.2.179\",\n \"podIPs\": [\n {\n \"ip\": \"10.244.2.179\"\n }\n ],\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2020-04-25T00:28:06Z\"\n }\n}\n" STEP: replace the image in the pod Apr 25 00:28:11.755: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config replace -f - --namespace=kubectl-1276' Apr 25 00:28:12.000: INFO: stderr: "" Apr 25 00:28:12.000: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n" STEP: verifying the pod e2e-test-httpd-pod has the right image docker.io/library/busybox:1.29 [AfterEach] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1459 Apr 25 00:28:12.006: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete pods e2e-test-httpd-pod --namespace=kubectl-1276' Apr 25 00:28:14.760: INFO: stderr: "" Apr 25 00:28:14.760: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:28:14.760: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-1276" for this suite. • [SLOW TEST:8.333 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1450 should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance]","total":275,"completed":168,"skipped":3177,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:28:14.771: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test env composition Apr 25 00:28:14.822: INFO: Waiting up to 5m0s for pod "var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde" in namespace "var-expansion-1122" to be "Succeeded or Failed" Apr 25 00:28:14.826: INFO: Pod "var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde": Phase="Pending", Reason="", readiness=false. Elapsed: 4.176148ms Apr 25 00:28:16.831: INFO: Pod "var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009099765s Apr 25 00:28:18.857: INFO: Pod "var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034968569s STEP: Saw pod success Apr 25 00:28:18.857: INFO: Pod "var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde" satisfied condition "Succeeded or Failed" Apr 25 00:28:18.860: INFO: Trying to get logs from node latest-worker2 pod var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde container dapi-container: STEP: delete the pod Apr 25 00:28:18.939: INFO: Waiting for pod var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde to disappear Apr 25 00:28:18.946: INFO: Pod var-expansion-7b34d191-0435-4cf5-9636-c219f72a1fde no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:28:18.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-1122" for this suite. •{"msg":"PASSED [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":275,"completed":169,"skipped":3195,"failed":0} SSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:28:18.953: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:84 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:99 STEP: Creating service test in namespace statefulset-3589 [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Initializing watcher for selector baz=blah,foo=bar STEP: Creating stateful set ss in namespace statefulset-3589 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-3589 Apr 25 00:28:19.074: INFO: Found 0 stateful pods, waiting for 1 Apr 25 00:28:29.078: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod Apr 25 00:28:29.082: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:28:29.334: INFO: stderr: "I0425 00:28:29.223860 2259 log.go:172] (0xc00003abb0) (0xc00090e3c0) Create stream\nI0425 00:28:29.223915 2259 log.go:172] (0xc00003abb0) (0xc00090e3c0) Stream added, broadcasting: 1\nI0425 00:28:29.226994 2259 log.go:172] (0xc00003abb0) Reply frame received for 1\nI0425 00:28:29.227050 2259 log.go:172] (0xc00003abb0) (0xc000222aa0) Create stream\nI0425 00:28:29.227096 2259 log.go:172] (0xc00003abb0) (0xc000222aa0) Stream added, broadcasting: 3\nI0425 00:28:29.228109 2259 log.go:172] (0xc00003abb0) Reply frame received for 3\nI0425 00:28:29.228148 2259 log.go:172] (0xc00003abb0) (0xc0005a9680) Create stream\nI0425 00:28:29.228163 2259 log.go:172] (0xc00003abb0) (0xc0005a9680) Stream added, broadcasting: 5\nI0425 00:28:29.229233 2259 log.go:172] (0xc00003abb0) Reply frame received for 5\nI0425 00:28:29.298136 2259 log.go:172] (0xc00003abb0) Data frame received for 5\nI0425 00:28:29.298163 2259 log.go:172] (0xc0005a9680) (5) Data frame handling\nI0425 00:28:29.298178 2259 log.go:172] (0xc0005a9680) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:28:29.326220 2259 log.go:172] (0xc00003abb0) Data frame received for 3\nI0425 00:28:29.326244 2259 log.go:172] (0xc000222aa0) (3) Data frame handling\nI0425 00:28:29.326253 2259 log.go:172] (0xc000222aa0) (3) Data frame sent\nI0425 00:28:29.326258 2259 log.go:172] (0xc00003abb0) Data frame received for 3\nI0425 00:28:29.326263 2259 log.go:172] (0xc000222aa0) (3) Data frame handling\nI0425 00:28:29.326640 2259 log.go:172] (0xc00003abb0) Data frame received for 5\nI0425 00:28:29.326669 2259 log.go:172] (0xc0005a9680) (5) Data frame handling\nI0425 00:28:29.328573 2259 log.go:172] (0xc00003abb0) Data frame received for 1\nI0425 00:28:29.328583 2259 log.go:172] (0xc00090e3c0) (1) Data frame handling\nI0425 00:28:29.328589 2259 log.go:172] (0xc00090e3c0) (1) Data frame sent\nI0425 00:28:29.328733 2259 log.go:172] (0xc00003abb0) (0xc00090e3c0) Stream removed, broadcasting: 1\nI0425 00:28:29.328866 2259 log.go:172] (0xc00003abb0) Go away received\nI0425 00:28:29.329266 2259 log.go:172] (0xc00003abb0) (0xc00090e3c0) Stream removed, broadcasting: 1\nI0425 00:28:29.329289 2259 log.go:172] (0xc00003abb0) (0xc000222aa0) Stream removed, broadcasting: 3\nI0425 00:28:29.329301 2259 log.go:172] (0xc00003abb0) (0xc0005a9680) Stream removed, broadcasting: 5\n" Apr 25 00:28:29.334: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:28:29.334: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:28:29.345: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Apr 25 00:28:39.350: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:28:39.350: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:28:39.390: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999574s Apr 25 00:28:40.395: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.996386394s Apr 25 00:28:41.400: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.991784916s Apr 25 00:28:42.404: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.987045034s Apr 25 00:28:43.409: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.982218724s Apr 25 00:28:44.413: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.977638339s Apr 25 00:28:45.418: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.973380776s Apr 25 00:28:46.422: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.968996653s Apr 25 00:28:47.426: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.964376651s Apr 25 00:28:48.430: INFO: Verifying statefulset ss doesn't scale past 1 for another 960.502275ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-3589 Apr 25 00:28:49.435: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:28:49.680: INFO: stderr: "I0425 00:28:49.577315 2281 log.go:172] (0xc000a0b4a0) (0xc0009de6e0) Create stream\nI0425 00:28:49.577375 2281 log.go:172] (0xc000a0b4a0) (0xc0009de6e0) Stream added, broadcasting: 1\nI0425 00:28:49.582494 2281 log.go:172] (0xc000a0b4a0) Reply frame received for 1\nI0425 00:28:49.582537 2281 log.go:172] (0xc000a0b4a0) (0xc0006975e0) Create stream\nI0425 00:28:49.582550 2281 log.go:172] (0xc000a0b4a0) (0xc0006975e0) Stream added, broadcasting: 3\nI0425 00:28:49.583263 2281 log.go:172] (0xc000a0b4a0) Reply frame received for 3\nI0425 00:28:49.583304 2281 log.go:172] (0xc000a0b4a0) (0xc0004d0a00) Create stream\nI0425 00:28:49.583319 2281 log.go:172] (0xc000a0b4a0) (0xc0004d0a00) Stream added, broadcasting: 5\nI0425 00:28:49.584066 2281 log.go:172] (0xc000a0b4a0) Reply frame received for 5\nI0425 00:28:49.672327 2281 log.go:172] (0xc000a0b4a0) Data frame received for 3\nI0425 00:28:49.672369 2281 log.go:172] (0xc0006975e0) (3) Data frame handling\nI0425 00:28:49.672390 2281 log.go:172] (0xc0006975e0) (3) Data frame sent\nI0425 00:28:49.672408 2281 log.go:172] (0xc000a0b4a0) Data frame received for 3\nI0425 00:28:49.672420 2281 log.go:172] (0xc0006975e0) (3) Data frame handling\nI0425 00:28:49.672435 2281 log.go:172] (0xc000a0b4a0) Data frame received for 5\nI0425 00:28:49.672448 2281 log.go:172] (0xc0004d0a00) (5) Data frame handling\nI0425 00:28:49.672459 2281 log.go:172] (0xc0004d0a00) (5) Data frame sent\nI0425 00:28:49.672470 2281 log.go:172] (0xc000a0b4a0) Data frame received for 5\nI0425 00:28:49.672480 2281 log.go:172] (0xc0004d0a00) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI0425 00:28:49.673852 2281 log.go:172] (0xc000a0b4a0) Data frame received for 1\nI0425 00:28:49.673976 2281 log.go:172] (0xc0009de6e0) (1) Data frame handling\nI0425 00:28:49.674019 2281 log.go:172] (0xc0009de6e0) (1) Data frame sent\nI0425 00:28:49.674053 2281 log.go:172] (0xc000a0b4a0) (0xc0009de6e0) Stream removed, broadcasting: 1\nI0425 00:28:49.674088 2281 log.go:172] (0xc000a0b4a0) Go away received\nI0425 00:28:49.674568 2281 log.go:172] (0xc000a0b4a0) (0xc0009de6e0) Stream removed, broadcasting: 1\nI0425 00:28:49.674614 2281 log.go:172] (0xc000a0b4a0) (0xc0006975e0) Stream removed, broadcasting: 3\nI0425 00:28:49.674632 2281 log.go:172] (0xc000a0b4a0) (0xc0004d0a00) Stream removed, broadcasting: 5\n" Apr 25 00:28:49.680: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:28:49.680: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:28:49.683: INFO: Found 1 stateful pods, waiting for 3 Apr 25 00:28:59.687: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:28:59.687: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:28:59.687: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Verifying that stateful set ss was scaled up in order STEP: Scale down will halt with unhealthy stateful pod Apr 25 00:28:59.693: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:28:59.932: INFO: stderr: "I0425 00:28:59.828368 2304 log.go:172] (0xc000b28000) (0xc000aa2000) Create stream\nI0425 00:28:59.828465 2304 log.go:172] (0xc000b28000) (0xc000aa2000) Stream added, broadcasting: 1\nI0425 00:28:59.830663 2304 log.go:172] (0xc000b28000) Reply frame received for 1\nI0425 00:28:59.830716 2304 log.go:172] (0xc000b28000) (0xc0005ba000) Create stream\nI0425 00:28:59.830736 2304 log.go:172] (0xc000b28000) (0xc0005ba000) Stream added, broadcasting: 3\nI0425 00:28:59.831909 2304 log.go:172] (0xc000b28000) Reply frame received for 3\nI0425 00:28:59.831952 2304 log.go:172] (0xc000b28000) (0xc0005f0000) Create stream\nI0425 00:28:59.831964 2304 log.go:172] (0xc000b28000) (0xc0005f0000) Stream added, broadcasting: 5\nI0425 00:28:59.832899 2304 log.go:172] (0xc000b28000) Reply frame received for 5\nI0425 00:28:59.922647 2304 log.go:172] (0xc000b28000) Data frame received for 5\nI0425 00:28:59.922668 2304 log.go:172] (0xc0005f0000) (5) Data frame handling\nI0425 00:28:59.922676 2304 log.go:172] (0xc0005f0000) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:28:59.922690 2304 log.go:172] (0xc000b28000) Data frame received for 3\nI0425 00:28:59.922697 2304 log.go:172] (0xc0005ba000) (3) Data frame handling\nI0425 00:28:59.922705 2304 log.go:172] (0xc0005ba000) (3) Data frame sent\nI0425 00:28:59.922709 2304 log.go:172] (0xc000b28000) Data frame received for 3\nI0425 00:28:59.922713 2304 log.go:172] (0xc0005ba000) (3) Data frame handling\nI0425 00:28:59.922835 2304 log.go:172] (0xc000b28000) Data frame received for 5\nI0425 00:28:59.922873 2304 log.go:172] (0xc0005f0000) (5) Data frame handling\nI0425 00:28:59.927866 2304 log.go:172] (0xc000b28000) Data frame received for 1\nI0425 00:28:59.927893 2304 log.go:172] (0xc000aa2000) (1) Data frame handling\nI0425 00:28:59.927940 2304 log.go:172] (0xc000aa2000) (1) Data frame sent\nI0425 00:28:59.927960 2304 log.go:172] (0xc000b28000) (0xc000aa2000) Stream removed, broadcasting: 1\nI0425 00:28:59.927976 2304 log.go:172] (0xc000b28000) Go away received\nI0425 00:28:59.928256 2304 log.go:172] (0xc000b28000) (0xc000aa2000) Stream removed, broadcasting: 1\nI0425 00:28:59.928270 2304 log.go:172] (0xc000b28000) (0xc0005ba000) Stream removed, broadcasting: 3\nI0425 00:28:59.928277 2304 log.go:172] (0xc000b28000) (0xc0005f0000) Stream removed, broadcasting: 5\n" Apr 25 00:28:59.932: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:28:59.932: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:28:59.932: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:29:00.162: INFO: stderr: "I0425 00:29:00.063866 2326 log.go:172] (0xc000a004d0) (0xc0009aaaa0) Create stream\nI0425 00:29:00.063939 2326 log.go:172] (0xc000a004d0) (0xc0009aaaa0) Stream added, broadcasting: 1\nI0425 00:29:00.067002 2326 log.go:172] (0xc000a004d0) Reply frame received for 1\nI0425 00:29:00.067054 2326 log.go:172] (0xc000a004d0) (0xc000ad0280) Create stream\nI0425 00:29:00.067070 2326 log.go:172] (0xc000a004d0) (0xc000ad0280) Stream added, broadcasting: 3\nI0425 00:29:00.068223 2326 log.go:172] (0xc000a004d0) Reply frame received for 3\nI0425 00:29:00.068260 2326 log.go:172] (0xc000a004d0) (0xc0009aab40) Create stream\nI0425 00:29:00.068288 2326 log.go:172] (0xc000a004d0) (0xc0009aab40) Stream added, broadcasting: 5\nI0425 00:29:00.069552 2326 log.go:172] (0xc000a004d0) Reply frame received for 5\nI0425 00:29:00.125433 2326 log.go:172] (0xc000a004d0) Data frame received for 5\nI0425 00:29:00.125479 2326 log.go:172] (0xc0009aab40) (5) Data frame handling\nI0425 00:29:00.125519 2326 log.go:172] (0xc0009aab40) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:29:00.154360 2326 log.go:172] (0xc000a004d0) Data frame received for 3\nI0425 00:29:00.154398 2326 log.go:172] (0xc000ad0280) (3) Data frame handling\nI0425 00:29:00.154427 2326 log.go:172] (0xc000ad0280) (3) Data frame sent\nI0425 00:29:00.154443 2326 log.go:172] (0xc000a004d0) Data frame received for 3\nI0425 00:29:00.154468 2326 log.go:172] (0xc000a004d0) Data frame received for 5\nI0425 00:29:00.154487 2326 log.go:172] (0xc0009aab40) (5) Data frame handling\nI0425 00:29:00.154509 2326 log.go:172] (0xc000ad0280) (3) Data frame handling\nI0425 00:29:00.156739 2326 log.go:172] (0xc000a004d0) Data frame received for 1\nI0425 00:29:00.156750 2326 log.go:172] (0xc0009aaaa0) (1) Data frame handling\nI0425 00:29:00.156773 2326 log.go:172] (0xc0009aaaa0) (1) Data frame sent\nI0425 00:29:00.156783 2326 log.go:172] (0xc000a004d0) (0xc0009aaaa0) Stream removed, broadcasting: 1\nI0425 00:29:00.156931 2326 log.go:172] (0xc000a004d0) Go away received\nI0425 00:29:00.157106 2326 log.go:172] (0xc000a004d0) (0xc0009aaaa0) Stream removed, broadcasting: 1\nI0425 00:29:00.157221 2326 log.go:172] (0xc000a004d0) (0xc000ad0280) Stream removed, broadcasting: 3\nI0425 00:29:00.157232 2326 log.go:172] (0xc000a004d0) (0xc0009aab40) Stream removed, broadcasting: 5\n" Apr 25 00:29:00.162: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:29:00.162: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:29:00.162: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Apr 25 00:29:00.396: INFO: stderr: "I0425 00:29:00.287949 2346 log.go:172] (0xc000a546e0) (0xc0007f54a0) Create stream\nI0425 00:29:00.288028 2346 log.go:172] (0xc000a546e0) (0xc0007f54a0) Stream added, broadcasting: 1\nI0425 00:29:00.291189 2346 log.go:172] (0xc000a546e0) Reply frame received for 1\nI0425 00:29:00.291245 2346 log.go:172] (0xc000a546e0) (0xc00066b540) Create stream\nI0425 00:29:00.291265 2346 log.go:172] (0xc000a546e0) (0xc00066b540) Stream added, broadcasting: 3\nI0425 00:29:00.292343 2346 log.go:172] (0xc000a546e0) Reply frame received for 3\nI0425 00:29:00.292384 2346 log.go:172] (0xc000a546e0) (0xc0007f5540) Create stream\nI0425 00:29:00.292401 2346 log.go:172] (0xc000a546e0) (0xc0007f5540) Stream added, broadcasting: 5\nI0425 00:29:00.293570 2346 log.go:172] (0xc000a546e0) Reply frame received for 5\nI0425 00:29:00.354629 2346 log.go:172] (0xc000a546e0) Data frame received for 5\nI0425 00:29:00.354669 2346 log.go:172] (0xc0007f5540) (5) Data frame handling\nI0425 00:29:00.354701 2346 log.go:172] (0xc0007f5540) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI0425 00:29:00.388180 2346 log.go:172] (0xc000a546e0) Data frame received for 3\nI0425 00:29:00.388220 2346 log.go:172] (0xc00066b540) (3) Data frame handling\nI0425 00:29:00.388249 2346 log.go:172] (0xc00066b540) (3) Data frame sent\nI0425 00:29:00.389357 2346 log.go:172] (0xc000a546e0) Data frame received for 3\nI0425 00:29:00.389381 2346 log.go:172] (0xc000a546e0) Data frame received for 5\nI0425 00:29:00.389406 2346 log.go:172] (0xc0007f5540) (5) Data frame handling\nI0425 00:29:00.389424 2346 log.go:172] (0xc00066b540) (3) Data frame handling\nI0425 00:29:00.390898 2346 log.go:172] (0xc000a546e0) Data frame received for 1\nI0425 00:29:00.390927 2346 log.go:172] (0xc0007f54a0) (1) Data frame handling\nI0425 00:29:00.390947 2346 log.go:172] (0xc0007f54a0) (1) Data frame sent\nI0425 00:29:00.390965 2346 log.go:172] (0xc000a546e0) (0xc0007f54a0) Stream removed, broadcasting: 1\nI0425 00:29:00.391372 2346 log.go:172] (0xc000a546e0) (0xc0007f54a0) Stream removed, broadcasting: 1\nI0425 00:29:00.391407 2346 log.go:172] (0xc000a546e0) (0xc00066b540) Stream removed, broadcasting: 3\nI0425 00:29:00.391592 2346 log.go:172] (0xc000a546e0) (0xc0007f5540) Stream removed, broadcasting: 5\n" Apr 25 00:29:00.396: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Apr 25 00:29:00.396: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Apr 25 00:29:00.396: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:29:00.399: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 Apr 25 00:29:10.408: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:29:10.408: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:29:10.408: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Apr 25 00:29:10.440: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999369s Apr 25 00:29:11.471: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.974918375s Apr 25 00:29:12.476: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.943704235s Apr 25 00:29:13.481: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.93873864s Apr 25 00:29:14.486: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.933850316s Apr 25 00:29:15.496: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.929340355s Apr 25 00:29:16.501: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.919120396s Apr 25 00:29:17.532: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.914302178s Apr 25 00:29:18.536: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.882940703s Apr 25 00:29:19.540: INFO: Verifying statefulset ss doesn't scale past 3 for another 878.852721ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-3589 Apr 25 00:29:20.545: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:29:20.799: INFO: stderr: "I0425 00:29:20.692955 2367 log.go:172] (0xc000b45080) (0xc0008ce500) Create stream\nI0425 00:29:20.693033 2367 log.go:172] (0xc000b45080) (0xc0008ce500) Stream added, broadcasting: 1\nI0425 00:29:20.695578 2367 log.go:172] (0xc000b45080) Reply frame received for 1\nI0425 00:29:20.695614 2367 log.go:172] (0xc000b45080) (0xc000ac4320) Create stream\nI0425 00:29:20.695623 2367 log.go:172] (0xc000b45080) (0xc000ac4320) Stream added, broadcasting: 3\nI0425 00:29:20.696505 2367 log.go:172] (0xc000b45080) Reply frame received for 3\nI0425 00:29:20.696540 2367 log.go:172] (0xc000b45080) (0xc000c24140) Create stream\nI0425 00:29:20.696550 2367 log.go:172] (0xc000b45080) (0xc000c24140) Stream added, broadcasting: 5\nI0425 00:29:20.697547 2367 log.go:172] (0xc000b45080) Reply frame received for 5\nI0425 00:29:20.791176 2367 log.go:172] (0xc000b45080) Data frame received for 5\nI0425 00:29:20.791220 2367 log.go:172] (0xc000c24140) (5) Data frame handling\nI0425 00:29:20.791236 2367 log.go:172] (0xc000c24140) (5) Data frame sent\nI0425 00:29:20.791248 2367 log.go:172] (0xc000b45080) Data frame received for 5\nI0425 00:29:20.791270 2367 log.go:172] (0xc000c24140) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI0425 00:29:20.791309 2367 log.go:172] (0xc000b45080) Data frame received for 3\nI0425 00:29:20.791354 2367 log.go:172] (0xc000ac4320) (3) Data frame handling\nI0425 00:29:20.791393 2367 log.go:172] (0xc000ac4320) (3) Data frame sent\nI0425 00:29:20.791422 2367 log.go:172] (0xc000b45080) Data frame received for 3\nI0425 00:29:20.791438 2367 log.go:172] (0xc000ac4320) (3) Data frame handling\nI0425 00:29:20.792848 2367 log.go:172] (0xc000b45080) Data frame received for 1\nI0425 00:29:20.792879 2367 log.go:172] (0xc0008ce500) (1) Data frame handling\nI0425 00:29:20.792896 2367 log.go:172] (0xc0008ce500) (1) Data frame sent\nI0425 00:29:20.792912 2367 log.go:172] (0xc000b45080) (0xc0008ce500) Stream removed, broadcasting: 1\nI0425 00:29:20.793013 2367 log.go:172] (0xc000b45080) Go away received\nI0425 00:29:20.793403 2367 log.go:172] (0xc000b45080) (0xc0008ce500) Stream removed, broadcasting: 1\nI0425 00:29:20.793421 2367 log.go:172] (0xc000b45080) (0xc000ac4320) Stream removed, broadcasting: 3\nI0425 00:29:20.793432 2367 log.go:172] (0xc000b45080) (0xc000c24140) Stream removed, broadcasting: 5\n" Apr 25 00:29:20.799: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:29:20.799: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:29:20.799: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:29:21.006: INFO: stderr: "I0425 00:29:20.927927 2387 log.go:172] (0xc000afc2c0) (0xc000aec140) Create stream\nI0425 00:29:20.928006 2387 log.go:172] (0xc000afc2c0) (0xc000aec140) Stream added, broadcasting: 1\nI0425 00:29:20.932072 2387 log.go:172] (0xc000afc2c0) Reply frame received for 1\nI0425 00:29:20.932227 2387 log.go:172] (0xc000afc2c0) (0xc000938000) Create stream\nI0425 00:29:20.932372 2387 log.go:172] (0xc000afc2c0) (0xc000938000) Stream added, broadcasting: 3\nI0425 00:29:20.934420 2387 log.go:172] (0xc000afc2c0) Reply frame received for 3\nI0425 00:29:20.935232 2387 log.go:172] (0xc000afc2c0) (0xc0008d4000) Create stream\nI0425 00:29:20.935264 2387 log.go:172] (0xc000afc2c0) (0xc0008d4000) Stream added, broadcasting: 5\nI0425 00:29:20.936579 2387 log.go:172] (0xc000afc2c0) Reply frame received for 5\nI0425 00:29:20.999915 2387 log.go:172] (0xc000afc2c0) Data frame received for 3\nI0425 00:29:20.999982 2387 log.go:172] (0xc000938000) (3) Data frame handling\nI0425 00:29:21.000010 2387 log.go:172] (0xc000938000) (3) Data frame sent\nI0425 00:29:21.000083 2387 log.go:172] (0xc000afc2c0) Data frame received for 5\nI0425 00:29:21.000123 2387 log.go:172] (0xc000afc2c0) Data frame received for 3\nI0425 00:29:21.000159 2387 log.go:172] (0xc000938000) (3) Data frame handling\nI0425 00:29:21.000195 2387 log.go:172] (0xc0008d4000) (5) Data frame handling\nI0425 00:29:21.000222 2387 log.go:172] (0xc0008d4000) (5) Data frame sent\nI0425 00:29:21.000249 2387 log.go:172] (0xc000afc2c0) Data frame received for 5\nI0425 00:29:21.000270 2387 log.go:172] (0xc0008d4000) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI0425 00:29:21.001827 2387 log.go:172] (0xc000afc2c0) Data frame received for 1\nI0425 00:29:21.001857 2387 log.go:172] (0xc000aec140) (1) Data frame handling\nI0425 00:29:21.001879 2387 log.go:172] (0xc000aec140) (1) Data frame sent\nI0425 00:29:21.001901 2387 log.go:172] (0xc000afc2c0) (0xc000aec140) Stream removed, broadcasting: 1\nI0425 00:29:21.001953 2387 log.go:172] (0xc000afc2c0) Go away received\nI0425 00:29:21.002452 2387 log.go:172] (0xc000afc2c0) (0xc000aec140) Stream removed, broadcasting: 1\nI0425 00:29:21.002477 2387 log.go:172] (0xc000afc2c0) (0xc000938000) Stream removed, broadcasting: 3\nI0425 00:29:21.002489 2387 log.go:172] (0xc000afc2c0) (0xc0008d4000) Stream removed, broadcasting: 5\n" Apr 25 00:29:21.006: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:29:21.006: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:29:21.006: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=statefulset-3589 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Apr 25 00:29:21.209: INFO: stderr: "I0425 00:29:21.135414 2407 log.go:172] (0xc000a264d0) (0xc0006b9680) Create stream\nI0425 00:29:21.135469 2407 log.go:172] (0xc000a264d0) (0xc0006b9680) Stream added, broadcasting: 1\nI0425 00:29:21.138235 2407 log.go:172] (0xc000a264d0) Reply frame received for 1\nI0425 00:29:21.138299 2407 log.go:172] (0xc000a264d0) (0xc00044cb40) Create stream\nI0425 00:29:21.138318 2407 log.go:172] (0xc000a264d0) (0xc00044cb40) Stream added, broadcasting: 3\nI0425 00:29:21.139569 2407 log.go:172] (0xc000a264d0) Reply frame received for 3\nI0425 00:29:21.139601 2407 log.go:172] (0xc000a264d0) (0xc00044cbe0) Create stream\nI0425 00:29:21.139616 2407 log.go:172] (0xc000a264d0) (0xc00044cbe0) Stream added, broadcasting: 5\nI0425 00:29:21.140758 2407 log.go:172] (0xc000a264d0) Reply frame received for 5\nI0425 00:29:21.202203 2407 log.go:172] (0xc000a264d0) Data frame received for 5\nI0425 00:29:21.202227 2407 log.go:172] (0xc00044cbe0) (5) Data frame handling\nI0425 00:29:21.202237 2407 log.go:172] (0xc00044cbe0) (5) Data frame sent\nI0425 00:29:21.202245 2407 log.go:172] (0xc000a264d0) Data frame received for 5\nI0425 00:29:21.202251 2407 log.go:172] (0xc00044cbe0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI0425 00:29:21.202283 2407 log.go:172] (0xc000a264d0) Data frame received for 3\nI0425 00:29:21.202320 2407 log.go:172] (0xc00044cb40) (3) Data frame handling\nI0425 00:29:21.202362 2407 log.go:172] (0xc00044cb40) (3) Data frame sent\nI0425 00:29:21.202384 2407 log.go:172] (0xc000a264d0) Data frame received for 3\nI0425 00:29:21.202399 2407 log.go:172] (0xc00044cb40) (3) Data frame handling\nI0425 00:29:21.204399 2407 log.go:172] (0xc000a264d0) Data frame received for 1\nI0425 00:29:21.204420 2407 log.go:172] (0xc0006b9680) (1) Data frame handling\nI0425 00:29:21.204431 2407 log.go:172] (0xc0006b9680) (1) Data frame sent\nI0425 00:29:21.204446 2407 log.go:172] (0xc000a264d0) (0xc0006b9680) Stream removed, broadcasting: 1\nI0425 00:29:21.204466 2407 log.go:172] (0xc000a264d0) Go away received\nI0425 00:29:21.205016 2407 log.go:172] (0xc000a264d0) (0xc0006b9680) Stream removed, broadcasting: 1\nI0425 00:29:21.205049 2407 log.go:172] (0xc000a264d0) (0xc00044cb40) Stream removed, broadcasting: 3\nI0425 00:29:21.205063 2407 log.go:172] (0xc000a264d0) (0xc00044cbe0) Stream removed, broadcasting: 5\n" Apr 25 00:29:21.209: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Apr 25 00:29:21.209: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Apr 25 00:29:21.210: INFO: Scaling statefulset ss to 0 STEP: Verifying that stateful set ss was scaled down in reverse order [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:110 Apr 25 00:29:41.226: INFO: Deleting all statefulset in ns statefulset-3589 Apr 25 00:29:41.228: INFO: Scaling statefulset ss to 0 Apr 25 00:29:41.235: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:29:41.237: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:29:41.253: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-3589" for this suite. • [SLOW TEST:82.309 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","total":275,"completed":170,"skipped":3204,"failed":0} SS ------------------------------ [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:29:41.262: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: starting a background goroutine to produce watch events STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:29:45.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-2615" for this suite. •{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":275,"completed":171,"skipped":3206,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:29:45.939: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Starting the proxy Apr 25 00:29:45.991: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config proxy --unix-socket=/tmp/kubectl-proxy-unix738981513/test' STEP: retrieving proxy /api/ output [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:29:46.058: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-1855" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance]","total":275,"completed":172,"skipped":3237,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:29:46.066: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153 [It] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod Apr 25 00:29:46.117: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:29:53.839: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-6089" for this suite. • [SLOW TEST:7.782 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","total":275,"completed":173,"skipped":3255,"failed":0} SSSS ------------------------------ [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:29:53.848: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap that has name configmap-test-emptyKey-721af4b0-7427-4127-8b3f-1dac80337d39 [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:29:53.927: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8489" for this suite. •{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","total":275,"completed":174,"skipped":3259,"failed":0} SSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:29:53.977: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0644 on tmpfs Apr 25 00:29:54.076: INFO: Waiting up to 5m0s for pod "pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9" in namespace "emptydir-9515" to be "Succeeded or Failed" Apr 25 00:29:54.113: INFO: Pod "pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9": Phase="Pending", Reason="", readiness=false. Elapsed: 37.08299ms Apr 25 00:29:56.117: INFO: Pod "pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.040755169s Apr 25 00:29:58.120: INFO: Pod "pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044565653s STEP: Saw pod success Apr 25 00:29:58.120: INFO: Pod "pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9" satisfied condition "Succeeded or Failed" Apr 25 00:29:58.123: INFO: Trying to get logs from node latest-worker2 pod pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9 container test-container: STEP: delete the pod Apr 25 00:29:58.152: INFO: Waiting for pod pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9 to disappear Apr 25 00:29:58.199: INFO: Pod pod-0a4ef81a-94ca-4920-8f75-b47d456f9ac9 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:29:58.199: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-9515" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":175,"skipped":3265,"failed":0} ------------------------------ [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:29:58.250: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename hostpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 [It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test hostPath mode Apr 25 00:29:58.297: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-9588" to be "Succeeded or Failed" Apr 25 00:29:58.318: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 20.477287ms Apr 25 00:30:00.376: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.078942895s Apr 25 00:30:02.381: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 4.083355719s Apr 25 00:30:04.384: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.0866089s STEP: Saw pod success Apr 25 00:30:04.384: INFO: Pod "pod-host-path-test" satisfied condition "Succeeded or Failed" Apr 25 00:30:04.387: INFO: Trying to get logs from node latest-worker2 pod pod-host-path-test container test-container-1: STEP: delete the pod Apr 25 00:30:04.422: INFO: Waiting for pod pod-host-path-test to disappear Apr 25 00:30:04.433: INFO: Pod pod-host-path-test no longer exists [AfterEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:30:04.433: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "hostpath-9588" for this suite. • [SLOW TEST:6.191 seconds] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":176,"skipped":3265,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:30:04.442: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-volume-map-e8d2ee5f-55ca-4038-909c-38fb9ba5cc0c STEP: Creating a pod to test consume configMaps Apr 25 00:30:04.537: INFO: Waiting up to 5m0s for pod "pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05" in namespace "configmap-2066" to be "Succeeded or Failed" Apr 25 00:30:04.541: INFO: Pod "pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05": Phase="Pending", Reason="", readiness=false. Elapsed: 3.641821ms Apr 25 00:30:06.544: INFO: Pod "pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006829455s Apr 25 00:30:08.548: INFO: Pod "pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011196916s STEP: Saw pod success Apr 25 00:30:08.548: INFO: Pod "pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05" satisfied condition "Succeeded or Failed" Apr 25 00:30:08.552: INFO: Trying to get logs from node latest-worker2 pod pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05 container configmap-volume-test: STEP: delete the pod Apr 25 00:30:08.573: INFO: Waiting for pod pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05 to disappear Apr 25 00:30:08.583: INFO: Pod pod-configmaps-58aba2da-608e-4609-892d-b2b795dbae05 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:30:08.583: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2066" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":177,"skipped":3286,"failed":0} SSSSSS ------------------------------ [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:30:08.591: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating projection with configMap that has name projected-configmap-test-upd-1adc9026-7231-494c-a65c-20bb78a2876e STEP: Creating the pod STEP: Updating configmap projected-configmap-test-upd-1adc9026-7231-494c-a65c-20bb78a2876e STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:30:16.854: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3075" for this suite. • [SLOW TEST:8.271 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:36 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":275,"completed":178,"skipped":3292,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:30:16.863: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods STEP: Gathering metrics W0425 00:30:57.027539 8 metrics_grabber.go:84] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 25 00:30:57.027: INFO: For apiserver_request_total: For apiserver_request_latency_seconds: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:30:57.027: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-2560" for this suite. • [SLOW TEST:40.171 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","total":275,"completed":179,"skipped":3300,"failed":0} SSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:30:57.034: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0777 on node default medium Apr 25 00:30:57.084: INFO: Waiting up to 5m0s for pod "pod-064df866-b1ab-4408-90dc-3ba89b433f2e" in namespace "emptydir-5173" to be "Succeeded or Failed" Apr 25 00:30:57.087: INFO: Pod "pod-064df866-b1ab-4408-90dc-3ba89b433f2e": Phase="Pending", Reason="", readiness=false. Elapsed: 3.609463ms Apr 25 00:30:59.204: INFO: Pod "pod-064df866-b1ab-4408-90dc-3ba89b433f2e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.120467161s Apr 25 00:31:01.208: INFO: Pod "pod-064df866-b1ab-4408-90dc-3ba89b433f2e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.124879974s STEP: Saw pod success Apr 25 00:31:01.209: INFO: Pod "pod-064df866-b1ab-4408-90dc-3ba89b433f2e" satisfied condition "Succeeded or Failed" Apr 25 00:31:01.212: INFO: Trying to get logs from node latest-worker pod pod-064df866-b1ab-4408-90dc-3ba89b433f2e container test-container: STEP: delete the pod Apr 25 00:31:01.301: INFO: Waiting for pod pod-064df866-b1ab-4408-90dc-3ba89b433f2e to disappear Apr 25 00:31:01.345: INFO: Pod pod-064df866-b1ab-4408-90dc-3ba89b433f2e no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:31:01.345: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-5173" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":180,"skipped":3307,"failed":0} ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:31:01.426: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Performing setup for networking test in namespace pod-network-test-7322 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 25 00:31:01.466: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Apr 25 00:31:01.514: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:31:03.768: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:31:05.517: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:31:07.529: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:09.519: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:11.519: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:13.518: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:15.519: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:17.518: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:19.518: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:21.518: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:23.518: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:25.533: INFO: The status of Pod netserver-0 is Running (Ready = true) Apr 25 00:31:25.538: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Apr 25 00:31:29.584: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.190:8080/dial?request=hostname&protocol=http&host=10.244.2.189&port=8080&tries=1'] Namespace:pod-network-test-7322 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:31:29.584: INFO: >>> kubeConfig: /root/.kube/config I0425 00:31:29.675341 8 log.go:172] (0xc0017ba4d0) (0xc0014417c0) Create stream I0425 00:31:29.675376 8 log.go:172] (0xc0017ba4d0) (0xc0014417c0) Stream added, broadcasting: 1 I0425 00:31:29.677454 8 log.go:172] (0xc0017ba4d0) Reply frame received for 1 I0425 00:31:29.677490 8 log.go:172] (0xc0017ba4d0) (0xc0010960a0) Create stream I0425 00:31:29.677507 8 log.go:172] (0xc0017ba4d0) (0xc0010960a0) Stream added, broadcasting: 3 I0425 00:31:29.678565 8 log.go:172] (0xc0017ba4d0) Reply frame received for 3 I0425 00:31:29.678612 8 log.go:172] (0xc0017ba4d0) (0xc001637900) Create stream I0425 00:31:29.678621 8 log.go:172] (0xc0017ba4d0) (0xc001637900) Stream added, broadcasting: 5 I0425 00:31:29.679550 8 log.go:172] (0xc0017ba4d0) Reply frame received for 5 I0425 00:31:29.754707 8 log.go:172] (0xc0017ba4d0) Data frame received for 3 I0425 00:31:29.754742 8 log.go:172] (0xc0010960a0) (3) Data frame handling I0425 00:31:29.754762 8 log.go:172] (0xc0010960a0) (3) Data frame sent I0425 00:31:29.755518 8 log.go:172] (0xc0017ba4d0) Data frame received for 3 I0425 00:31:29.755565 8 log.go:172] (0xc0010960a0) (3) Data frame handling I0425 00:31:29.755599 8 log.go:172] (0xc0017ba4d0) Data frame received for 5 I0425 00:31:29.755620 8 log.go:172] (0xc001637900) (5) Data frame handling I0425 00:31:29.757377 8 log.go:172] (0xc0017ba4d0) Data frame received for 1 I0425 00:31:29.757416 8 log.go:172] (0xc0014417c0) (1) Data frame handling I0425 00:31:29.757470 8 log.go:172] (0xc0014417c0) (1) Data frame sent I0425 00:31:29.757505 8 log.go:172] (0xc0017ba4d0) (0xc0014417c0) Stream removed, broadcasting: 1 I0425 00:31:29.757534 8 log.go:172] (0xc0017ba4d0) Go away received I0425 00:31:29.757637 8 log.go:172] (0xc0017ba4d0) (0xc0014417c0) Stream removed, broadcasting: 1 I0425 00:31:29.757659 8 log.go:172] (0xc0017ba4d0) (0xc0010960a0) Stream removed, broadcasting: 3 I0425 00:31:29.757666 8 log.go:172] (0xc0017ba4d0) (0xc001637900) Stream removed, broadcasting: 5 Apr 25 00:31:29.757: INFO: Waiting for responses: map[] Apr 25 00:31:29.761: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.190:8080/dial?request=hostname&protocol=http&host=10.244.1.232&port=8080&tries=1'] Namespace:pod-network-test-7322 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:31:29.761: INFO: >>> kubeConfig: /root/.kube/config I0425 00:31:29.790643 8 log.go:172] (0xc0029f8420) (0xc001637ea0) Create stream I0425 00:31:29.790670 8 log.go:172] (0xc0029f8420) (0xc001637ea0) Stream added, broadcasting: 1 I0425 00:31:29.792953 8 log.go:172] (0xc0029f8420) Reply frame received for 1 I0425 00:31:29.793007 8 log.go:172] (0xc0029f8420) (0xc001d730e0) Create stream I0425 00:31:29.793032 8 log.go:172] (0xc0029f8420) (0xc001d730e0) Stream added, broadcasting: 3 I0425 00:31:29.794434 8 log.go:172] (0xc0029f8420) Reply frame received for 3 I0425 00:31:29.794475 8 log.go:172] (0xc0029f8420) (0xc001e60000) Create stream I0425 00:31:29.794499 8 log.go:172] (0xc0029f8420) (0xc001e60000) Stream added, broadcasting: 5 I0425 00:31:29.795872 8 log.go:172] (0xc0029f8420) Reply frame received for 5 I0425 00:31:29.856825 8 log.go:172] (0xc0029f8420) Data frame received for 3 I0425 00:31:29.856852 8 log.go:172] (0xc001d730e0) (3) Data frame handling I0425 00:31:29.856874 8 log.go:172] (0xc001d730e0) (3) Data frame sent I0425 00:31:29.857282 8 log.go:172] (0xc0029f8420) Data frame received for 5 I0425 00:31:29.857315 8 log.go:172] (0xc0029f8420) Data frame received for 3 I0425 00:31:29.857350 8 log.go:172] (0xc001d730e0) (3) Data frame handling I0425 00:31:29.857374 8 log.go:172] (0xc001e60000) (5) Data frame handling I0425 00:31:29.858755 8 log.go:172] (0xc0029f8420) Data frame received for 1 I0425 00:31:29.858782 8 log.go:172] (0xc001637ea0) (1) Data frame handling I0425 00:31:29.858817 8 log.go:172] (0xc001637ea0) (1) Data frame sent I0425 00:31:29.858841 8 log.go:172] (0xc0029f8420) (0xc001637ea0) Stream removed, broadcasting: 1 I0425 00:31:29.858910 8 log.go:172] (0xc0029f8420) (0xc001637ea0) Stream removed, broadcasting: 1 I0425 00:31:29.858939 8 log.go:172] (0xc0029f8420) (0xc001d730e0) Stream removed, broadcasting: 3 I0425 00:31:29.859053 8 log.go:172] (0xc0029f8420) Go away received I0425 00:31:29.859113 8 log.go:172] (0xc0029f8420) (0xc001e60000) Stream removed, broadcasting: 5 Apr 25 00:31:29.859: INFO: Waiting for responses: map[] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:31:29.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-7322" for this suite. • [SLOW TEST:28.453 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for intra-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]","total":275,"completed":181,"skipped":3307,"failed":0} SS ------------------------------ [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:31:29.880: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:31:29.964: INFO: Waiting up to 5m0s for pod "downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145" in namespace "downward-api-8396" to be "Succeeded or Failed" Apr 25 00:31:30.018: INFO: Pod "downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145": Phase="Pending", Reason="", readiness=false. Elapsed: 54.413835ms Apr 25 00:31:32.022: INFO: Pod "downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145": Phase="Pending", Reason="", readiness=false. Elapsed: 2.058165102s Apr 25 00:31:34.027: INFO: Pod "downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.062921286s STEP: Saw pod success Apr 25 00:31:34.027: INFO: Pod "downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145" satisfied condition "Succeeded or Failed" Apr 25 00:31:34.030: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145 container client-container: STEP: delete the pod Apr 25 00:31:34.059: INFO: Waiting for pod downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145 to disappear Apr 25 00:31:34.070: INFO: Pod downwardapi-volume-43b41c2a-5682-49d4-a309-90558963b145 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:31:34.070: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-8396" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":275,"completed":182,"skipped":3309,"failed":0} SSSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:31:34.076: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Performing setup for networking test in namespace pod-network-test-4178 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 25 00:31:34.152: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Apr 25 00:31:34.204: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:31:36.375: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:31:38.208: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:31:40.207: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:42.216: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:44.207: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:46.217: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:48.207: INFO: The status of Pod netserver-0 is Running (Ready = false) Apr 25 00:31:50.246: INFO: The status of Pod netserver-0 is Running (Ready = true) Apr 25 00:31:50.259: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Apr 25 00:31:54.314: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.2.191 8081 | grep -v '^\s*$'] Namespace:pod-network-test-4178 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:31:54.314: INFO: >>> kubeConfig: /root/.kube/config I0425 00:31:54.345486 8 log.go:172] (0xc0028de0b0) (0xc001261ea0) Create stream I0425 00:31:54.345515 8 log.go:172] (0xc0028de0b0) (0xc001261ea0) Stream added, broadcasting: 1 I0425 00:31:54.357723 8 log.go:172] (0xc0028de0b0) Reply frame received for 1 I0425 00:31:54.357760 8 log.go:172] (0xc0028de0b0) (0xc001e60dc0) Create stream I0425 00:31:54.357769 8 log.go:172] (0xc0028de0b0) (0xc001e60dc0) Stream added, broadcasting: 3 I0425 00:31:54.359054 8 log.go:172] (0xc0028de0b0) Reply frame received for 3 I0425 00:31:54.359119 8 log.go:172] (0xc0028de0b0) (0xc001096960) Create stream I0425 00:31:54.359147 8 log.go:172] (0xc0028de0b0) (0xc001096960) Stream added, broadcasting: 5 I0425 00:31:54.361979 8 log.go:172] (0xc0028de0b0) Reply frame received for 5 I0425 00:31:55.441270 8 log.go:172] (0xc0028de0b0) Data frame received for 3 I0425 00:31:55.441325 8 log.go:172] (0xc0028de0b0) Data frame received for 5 I0425 00:31:55.441370 8 log.go:172] (0xc001096960) (5) Data frame handling I0425 00:31:55.441427 8 log.go:172] (0xc001e60dc0) (3) Data frame handling I0425 00:31:55.441467 8 log.go:172] (0xc001e60dc0) (3) Data frame sent I0425 00:31:55.441483 8 log.go:172] (0xc0028de0b0) Data frame received for 3 I0425 00:31:55.441494 8 log.go:172] (0xc001e60dc0) (3) Data frame handling I0425 00:31:55.443786 8 log.go:172] (0xc0028de0b0) Data frame received for 1 I0425 00:31:55.443815 8 log.go:172] (0xc001261ea0) (1) Data frame handling I0425 00:31:55.443835 8 log.go:172] (0xc001261ea0) (1) Data frame sent I0425 00:31:55.443847 8 log.go:172] (0xc0028de0b0) (0xc001261ea0) Stream removed, broadcasting: 1 I0425 00:31:55.443863 8 log.go:172] (0xc0028de0b0) Go away received I0425 00:31:55.444036 8 log.go:172] (0xc0028de0b0) (0xc001261ea0) Stream removed, broadcasting: 1 I0425 00:31:55.444066 8 log.go:172] (0xc0028de0b0) (0xc001e60dc0) Stream removed, broadcasting: 3 I0425 00:31:55.444090 8 log.go:172] (0xc0028de0b0) (0xc001096960) Stream removed, broadcasting: 5 Apr 25 00:31:55.444: INFO: Found all expected endpoints: [netserver-0] Apr 25 00:31:55.447: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.1.234 8081 | grep -v '^\s*$'] Namespace:pod-network-test-4178 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:31:55.447: INFO: >>> kubeConfig: /root/.kube/config I0425 00:31:55.474558 8 log.go:172] (0xc002dc2580) (0xc001e615e0) Create stream I0425 00:31:55.474602 8 log.go:172] (0xc002dc2580) (0xc001e615e0) Stream added, broadcasting: 1 I0425 00:31:55.476666 8 log.go:172] (0xc002dc2580) Reply frame received for 1 I0425 00:31:55.476735 8 log.go:172] (0xc002dc2580) (0xc0028e2000) Create stream I0425 00:31:55.476758 8 log.go:172] (0xc002dc2580) (0xc0028e2000) Stream added, broadcasting: 3 I0425 00:31:55.477922 8 log.go:172] (0xc002dc2580) Reply frame received for 3 I0425 00:31:55.477962 8 log.go:172] (0xc002dc2580) (0xc0011ccfa0) Create stream I0425 00:31:55.477974 8 log.go:172] (0xc002dc2580) (0xc0011ccfa0) Stream added, broadcasting: 5 I0425 00:31:55.478814 8 log.go:172] (0xc002dc2580) Reply frame received for 5 I0425 00:31:56.532067 8 log.go:172] (0xc002dc2580) Data frame received for 3 I0425 00:31:56.532118 8 log.go:172] (0xc0028e2000) (3) Data frame handling I0425 00:31:56.532152 8 log.go:172] (0xc0028e2000) (3) Data frame sent I0425 00:31:56.532168 8 log.go:172] (0xc002dc2580) Data frame received for 3 I0425 00:31:56.532181 8 log.go:172] (0xc0028e2000) (3) Data frame handling I0425 00:31:56.532314 8 log.go:172] (0xc002dc2580) Data frame received for 5 I0425 00:31:56.532344 8 log.go:172] (0xc0011ccfa0) (5) Data frame handling I0425 00:31:56.534427 8 log.go:172] (0xc002dc2580) Data frame received for 1 I0425 00:31:56.534464 8 log.go:172] (0xc001e615e0) (1) Data frame handling I0425 00:31:56.534514 8 log.go:172] (0xc001e615e0) (1) Data frame sent I0425 00:31:56.534539 8 log.go:172] (0xc002dc2580) (0xc001e615e0) Stream removed, broadcasting: 1 I0425 00:31:56.534570 8 log.go:172] (0xc002dc2580) Go away received I0425 00:31:56.534745 8 log.go:172] (0xc002dc2580) (0xc001e615e0) Stream removed, broadcasting: 1 I0425 00:31:56.534814 8 log.go:172] (0xc002dc2580) (0xc0028e2000) Stream removed, broadcasting: 3 I0425 00:31:56.534875 8 log.go:172] (0xc002dc2580) (0xc0011ccfa0) Stream removed, broadcasting: 5 Apr 25 00:31:56.534: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:31:56.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-4178" for this suite. • [SLOW TEST:22.466 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":183,"skipped":3320,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:31:56.542: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:31:57.258: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:31:59.268: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371517, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371517, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371517, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371517, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:32:02.305: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] patching/updating a mutating webhook should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a mutating webhook configuration STEP: Updating a mutating webhook configuration's rules to not include the create operation STEP: Creating a configMap that should not be mutated STEP: Patching a mutating webhook configuration's rules to include the create operation STEP: Creating a configMap that should be mutated [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:32:02.720: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-5061" for this suite. STEP: Destroying namespace "webhook-5061-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:6.469 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 patching/updating a mutating webhook should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","total":275,"completed":184,"skipped":3341,"failed":0} SSSSSSSS ------------------------------ [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:32:03.012: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts STEP: Waiting for a default service account to be provisioned in namespace [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Setting up the test STEP: Creating hostNetwork=false pod STEP: Creating hostNetwork=true pod STEP: Running the test STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false Apr 25 00:32:15.354: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:15.354: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:15.389931 8 log.go:172] (0xc0028de9a0) (0xc001732c80) Create stream I0425 00:32:15.389958 8 log.go:172] (0xc0028de9a0) (0xc001732c80) Stream added, broadcasting: 1 I0425 00:32:15.391847 8 log.go:172] (0xc0028de9a0) Reply frame received for 1 I0425 00:32:15.391892 8 log.go:172] (0xc0028de9a0) (0xc001732dc0) Create stream I0425 00:32:15.391910 8 log.go:172] (0xc0028de9a0) (0xc001732dc0) Stream added, broadcasting: 3 I0425 00:32:15.392882 8 log.go:172] (0xc0028de9a0) Reply frame received for 3 I0425 00:32:15.392915 8 log.go:172] (0xc0028de9a0) (0xc0016fc460) Create stream I0425 00:32:15.392926 8 log.go:172] (0xc0028de9a0) (0xc0016fc460) Stream added, broadcasting: 5 I0425 00:32:15.393912 8 log.go:172] (0xc0028de9a0) Reply frame received for 5 I0425 00:32:15.453476 8 log.go:172] (0xc0028de9a0) Data frame received for 5 I0425 00:32:15.453516 8 log.go:172] (0xc0016fc460) (5) Data frame handling I0425 00:32:15.453538 8 log.go:172] (0xc0028de9a0) Data frame received for 3 I0425 00:32:15.453550 8 log.go:172] (0xc001732dc0) (3) Data frame handling I0425 00:32:15.453562 8 log.go:172] (0xc001732dc0) (3) Data frame sent I0425 00:32:15.453572 8 log.go:172] (0xc0028de9a0) Data frame received for 3 I0425 00:32:15.453581 8 log.go:172] (0xc001732dc0) (3) Data frame handling I0425 00:32:15.455067 8 log.go:172] (0xc0028de9a0) Data frame received for 1 I0425 00:32:15.455090 8 log.go:172] (0xc001732c80) (1) Data frame handling I0425 00:32:15.455104 8 log.go:172] (0xc001732c80) (1) Data frame sent I0425 00:32:15.455117 8 log.go:172] (0xc0028de9a0) (0xc001732c80) Stream removed, broadcasting: 1 I0425 00:32:15.455137 8 log.go:172] (0xc0028de9a0) Go away received I0425 00:32:15.455270 8 log.go:172] (0xc0028de9a0) (0xc001732c80) Stream removed, broadcasting: 1 I0425 00:32:15.455300 8 log.go:172] (0xc0028de9a0) (0xc001732dc0) Stream removed, broadcasting: 3 I0425 00:32:15.455316 8 log.go:172] (0xc0028de9a0) (0xc0016fc460) Stream removed, broadcasting: 5 Apr 25 00:32:15.455: INFO: Exec stderr: "" Apr 25 00:32:15.455: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:15.455: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:15.699873 8 log.go:172] (0xc00245fad0) (0xc001097360) Create stream I0425 00:32:15.699908 8 log.go:172] (0xc00245fad0) (0xc001097360) Stream added, broadcasting: 1 I0425 00:32:15.702086 8 log.go:172] (0xc00245fad0) Reply frame received for 1 I0425 00:32:15.702119 8 log.go:172] (0xc00245fad0) (0xc0010974a0) Create stream I0425 00:32:15.702131 8 log.go:172] (0xc00245fad0) (0xc0010974a0) Stream added, broadcasting: 3 I0425 00:32:15.703146 8 log.go:172] (0xc00245fad0) Reply frame received for 3 I0425 00:32:15.703178 8 log.go:172] (0xc00245fad0) (0xc001f565a0) Create stream I0425 00:32:15.703190 8 log.go:172] (0xc00245fad0) (0xc001f565a0) Stream added, broadcasting: 5 I0425 00:32:15.704535 8 log.go:172] (0xc00245fad0) Reply frame received for 5 I0425 00:32:15.768254 8 log.go:172] (0xc00245fad0) Data frame received for 5 I0425 00:32:15.768304 8 log.go:172] (0xc001f565a0) (5) Data frame handling I0425 00:32:15.768341 8 log.go:172] (0xc00245fad0) Data frame received for 3 I0425 00:32:15.768381 8 log.go:172] (0xc0010974a0) (3) Data frame handling I0425 00:32:15.768415 8 log.go:172] (0xc0010974a0) (3) Data frame sent I0425 00:32:15.768440 8 log.go:172] (0xc00245fad0) Data frame received for 3 I0425 00:32:15.768458 8 log.go:172] (0xc0010974a0) (3) Data frame handling I0425 00:32:15.770265 8 log.go:172] (0xc00245fad0) Data frame received for 1 I0425 00:32:15.770307 8 log.go:172] (0xc001097360) (1) Data frame handling I0425 00:32:15.770342 8 log.go:172] (0xc001097360) (1) Data frame sent I0425 00:32:15.770377 8 log.go:172] (0xc00245fad0) (0xc001097360) Stream removed, broadcasting: 1 I0425 00:32:15.770473 8 log.go:172] (0xc00245fad0) Go away received I0425 00:32:15.770529 8 log.go:172] (0xc00245fad0) (0xc001097360) Stream removed, broadcasting: 1 I0425 00:32:15.770571 8 log.go:172] (0xc00245fad0) (0xc0010974a0) Stream removed, broadcasting: 3 I0425 00:32:15.770591 8 log.go:172] (0xc00245fad0) (0xc001f565a0) Stream removed, broadcasting: 5 Apr 25 00:32:15.770: INFO: Exec stderr: "" Apr 25 00:32:15.770: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:15.770: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:15.808664 8 log.go:172] (0xc0029f8fd0) (0xc001f56780) Create stream I0425 00:32:15.808688 8 log.go:172] (0xc0029f8fd0) (0xc001f56780) Stream added, broadcasting: 1 I0425 00:32:15.811559 8 log.go:172] (0xc0029f8fd0) Reply frame received for 1 I0425 00:32:15.811611 8 log.go:172] (0xc0029f8fd0) (0xc001732fa0) Create stream I0425 00:32:15.811627 8 log.go:172] (0xc0029f8fd0) (0xc001732fa0) Stream added, broadcasting: 3 I0425 00:32:15.812665 8 log.go:172] (0xc0029f8fd0) Reply frame received for 3 I0425 00:32:15.812702 8 log.go:172] (0xc0029f8fd0) (0xc001f56960) Create stream I0425 00:32:15.812717 8 log.go:172] (0xc0029f8fd0) (0xc001f56960) Stream added, broadcasting: 5 I0425 00:32:15.813773 8 log.go:172] (0xc0029f8fd0) Reply frame received for 5 I0425 00:32:15.880832 8 log.go:172] (0xc0029f8fd0) Data frame received for 5 I0425 00:32:15.880890 8 log.go:172] (0xc0029f8fd0) Data frame received for 3 I0425 00:32:15.880916 8 log.go:172] (0xc001732fa0) (3) Data frame handling I0425 00:32:15.880934 8 log.go:172] (0xc001732fa0) (3) Data frame sent I0425 00:32:15.880945 8 log.go:172] (0xc0029f8fd0) Data frame received for 3 I0425 00:32:15.880952 8 log.go:172] (0xc001732fa0) (3) Data frame handling I0425 00:32:15.881004 8 log.go:172] (0xc001f56960) (5) Data frame handling I0425 00:32:15.883319 8 log.go:172] (0xc0029f8fd0) Data frame received for 1 I0425 00:32:15.883340 8 log.go:172] (0xc001f56780) (1) Data frame handling I0425 00:32:15.883369 8 log.go:172] (0xc001f56780) (1) Data frame sent I0425 00:32:15.883390 8 log.go:172] (0xc0029f8fd0) (0xc001f56780) Stream removed, broadcasting: 1 I0425 00:32:15.883425 8 log.go:172] (0xc0029f8fd0) Go away received I0425 00:32:15.883486 8 log.go:172] (0xc0029f8fd0) (0xc001f56780) Stream removed, broadcasting: 1 I0425 00:32:15.883503 8 log.go:172] (0xc0029f8fd0) (0xc001732fa0) Stream removed, broadcasting: 3 I0425 00:32:15.883519 8 log.go:172] (0xc0029f8fd0) (0xc001f56960) Stream removed, broadcasting: 5 Apr 25 00:32:15.883: INFO: Exec stderr: "" Apr 25 00:32:15.883: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:15.883: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:15.918362 8 log.go:172] (0xc0017bab00) (0xc001441d60) Create stream I0425 00:32:15.918384 8 log.go:172] (0xc0017bab00) (0xc001441d60) Stream added, broadcasting: 1 I0425 00:32:15.920060 8 log.go:172] (0xc0017bab00) Reply frame received for 1 I0425 00:32:15.920083 8 log.go:172] (0xc0017bab00) (0xc001097f40) Create stream I0425 00:32:15.920091 8 log.go:172] (0xc0017bab00) (0xc001097f40) Stream added, broadcasting: 3 I0425 00:32:15.920938 8 log.go:172] (0xc0017bab00) Reply frame received for 3 I0425 00:32:15.920963 8 log.go:172] (0xc0017bab00) (0xc001de20a0) Create stream I0425 00:32:15.920971 8 log.go:172] (0xc0017bab00) (0xc001de20a0) Stream added, broadcasting: 5 I0425 00:32:15.922126 8 log.go:172] (0xc0017bab00) Reply frame received for 5 I0425 00:32:15.984054 8 log.go:172] (0xc0017bab00) Data frame received for 5 I0425 00:32:15.984087 8 log.go:172] (0xc001de20a0) (5) Data frame handling I0425 00:32:15.984115 8 log.go:172] (0xc0017bab00) Data frame received for 3 I0425 00:32:15.984128 8 log.go:172] (0xc001097f40) (3) Data frame handling I0425 00:32:15.984141 8 log.go:172] (0xc001097f40) (3) Data frame sent I0425 00:32:15.984272 8 log.go:172] (0xc0017bab00) Data frame received for 3 I0425 00:32:15.984301 8 log.go:172] (0xc001097f40) (3) Data frame handling I0425 00:32:15.986184 8 log.go:172] (0xc0017bab00) Data frame received for 1 I0425 00:32:15.986205 8 log.go:172] (0xc001441d60) (1) Data frame handling I0425 00:32:15.986217 8 log.go:172] (0xc001441d60) (1) Data frame sent I0425 00:32:15.986227 8 log.go:172] (0xc0017bab00) (0xc001441d60) Stream removed, broadcasting: 1 I0425 00:32:15.986245 8 log.go:172] (0xc0017bab00) Go away received I0425 00:32:15.986339 8 log.go:172] (0xc0017bab00) (0xc001441d60) Stream removed, broadcasting: 1 I0425 00:32:15.986365 8 log.go:172] (0xc0017bab00) (0xc001097f40) Stream removed, broadcasting: 3 I0425 00:32:15.986383 8 log.go:172] (0xc0017bab00) (0xc001de20a0) Stream removed, broadcasting: 5 Apr 25 00:32:15.986: INFO: Exec stderr: "" STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount Apr 25 00:32:15.986: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:15.986: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:16.020810 8 log.go:172] (0xc002dc2d10) (0xc0016fc820) Create stream I0425 00:32:16.020832 8 log.go:172] (0xc002dc2d10) (0xc0016fc820) Stream added, broadcasting: 1 I0425 00:32:16.023327 8 log.go:172] (0xc002dc2d10) Reply frame received for 1 I0425 00:32:16.023387 8 log.go:172] (0xc002dc2d10) (0xc001de2280) Create stream I0425 00:32:16.023423 8 log.go:172] (0xc002dc2d10) (0xc001de2280) Stream added, broadcasting: 3 I0425 00:32:16.024299 8 log.go:172] (0xc002dc2d10) Reply frame received for 3 I0425 00:32:16.024335 8 log.go:172] (0xc002dc2d10) (0xc0022be0a0) Create stream I0425 00:32:16.024347 8 log.go:172] (0xc002dc2d10) (0xc0022be0a0) Stream added, broadcasting: 5 I0425 00:32:16.025453 8 log.go:172] (0xc002dc2d10) Reply frame received for 5 I0425 00:32:16.077033 8 log.go:172] (0xc002dc2d10) Data frame received for 5 I0425 00:32:16.077068 8 log.go:172] (0xc0022be0a0) (5) Data frame handling I0425 00:32:16.077089 8 log.go:172] (0xc002dc2d10) Data frame received for 3 I0425 00:32:16.077100 8 log.go:172] (0xc001de2280) (3) Data frame handling I0425 00:32:16.077218 8 log.go:172] (0xc001de2280) (3) Data frame sent I0425 00:32:16.077593 8 log.go:172] (0xc002dc2d10) Data frame received for 3 I0425 00:32:16.077617 8 log.go:172] (0xc001de2280) (3) Data frame handling I0425 00:32:16.079198 8 log.go:172] (0xc002dc2d10) Data frame received for 1 I0425 00:32:16.079213 8 log.go:172] (0xc0016fc820) (1) Data frame handling I0425 00:32:16.079228 8 log.go:172] (0xc0016fc820) (1) Data frame sent I0425 00:32:16.079239 8 log.go:172] (0xc002dc2d10) (0xc0016fc820) Stream removed, broadcasting: 1 I0425 00:32:16.079273 8 log.go:172] (0xc002dc2d10) Go away received I0425 00:32:16.079341 8 log.go:172] (0xc002dc2d10) (0xc0016fc820) Stream removed, broadcasting: 1 I0425 00:32:16.079353 8 log.go:172] (0xc002dc2d10) (0xc001de2280) Stream removed, broadcasting: 3 I0425 00:32:16.079364 8 log.go:172] (0xc002dc2d10) (0xc0022be0a0) Stream removed, broadcasting: 5 Apr 25 00:32:16.079: INFO: Exec stderr: "" Apr 25 00:32:16.079: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:16.079: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:16.114840 8 log.go:172] (0xc0017badc0) (0xc001de2d20) Create stream I0425 00:32:16.114869 8 log.go:172] (0xc0017badc0) (0xc001de2d20) Stream added, broadcasting: 1 I0425 00:32:16.117268 8 log.go:172] (0xc0017badc0) Reply frame received for 1 I0425 00:32:16.117338 8 log.go:172] (0xc0017badc0) (0xc001f56b40) Create stream I0425 00:32:16.117361 8 log.go:172] (0xc0017badc0) (0xc001f56b40) Stream added, broadcasting: 3 I0425 00:32:16.118263 8 log.go:172] (0xc0017badc0) Reply frame received for 3 I0425 00:32:16.118323 8 log.go:172] (0xc0017badc0) (0xc0022be140) Create stream I0425 00:32:16.118350 8 log.go:172] (0xc0017badc0) (0xc0022be140) Stream added, broadcasting: 5 I0425 00:32:16.119211 8 log.go:172] (0xc0017badc0) Reply frame received for 5 I0425 00:32:16.189848 8 log.go:172] (0xc0017badc0) Data frame received for 5 I0425 00:32:16.189881 8 log.go:172] (0xc0022be140) (5) Data frame handling I0425 00:32:16.189905 8 log.go:172] (0xc0017badc0) Data frame received for 3 I0425 00:32:16.189923 8 log.go:172] (0xc001f56b40) (3) Data frame handling I0425 00:32:16.189939 8 log.go:172] (0xc001f56b40) (3) Data frame sent I0425 00:32:16.189949 8 log.go:172] (0xc0017badc0) Data frame received for 3 I0425 00:32:16.189961 8 log.go:172] (0xc001f56b40) (3) Data frame handling I0425 00:32:16.191170 8 log.go:172] (0xc0017badc0) Data frame received for 1 I0425 00:32:16.191193 8 log.go:172] (0xc001de2d20) (1) Data frame handling I0425 00:32:16.191213 8 log.go:172] (0xc001de2d20) (1) Data frame sent I0425 00:32:16.191231 8 log.go:172] (0xc0017badc0) (0xc001de2d20) Stream removed, broadcasting: 1 I0425 00:32:16.191246 8 log.go:172] (0xc0017badc0) Go away received I0425 00:32:16.191378 8 log.go:172] (0xc0017badc0) (0xc001de2d20) Stream removed, broadcasting: 1 I0425 00:32:16.191404 8 log.go:172] (0xc0017badc0) (0xc001f56b40) Stream removed, broadcasting: 3 I0425 00:32:16.191423 8 log.go:172] (0xc0017badc0) (0xc0022be140) Stream removed, broadcasting: 5 Apr 25 00:32:16.191: INFO: Exec stderr: "" STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true Apr 25 00:32:16.191: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:16.191: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:16.228872 8 log.go:172] (0xc0017bb3f0) (0xc001de3040) Create stream I0425 00:32:16.228901 8 log.go:172] (0xc0017bb3f0) (0xc001de3040) Stream added, broadcasting: 1 I0425 00:32:16.231235 8 log.go:172] (0xc0017bb3f0) Reply frame received for 1 I0425 00:32:16.231283 8 log.go:172] (0xc0017bb3f0) (0xc0016fcc80) Create stream I0425 00:32:16.231300 8 log.go:172] (0xc0017bb3f0) (0xc0016fcc80) Stream added, broadcasting: 3 I0425 00:32:16.232280 8 log.go:172] (0xc0017bb3f0) Reply frame received for 3 I0425 00:32:16.232313 8 log.go:172] (0xc0017bb3f0) (0xc0016fcdc0) Create stream I0425 00:32:16.232326 8 log.go:172] (0xc0017bb3f0) (0xc0016fcdc0) Stream added, broadcasting: 5 I0425 00:32:16.233300 8 log.go:172] (0xc0017bb3f0) Reply frame received for 5 I0425 00:32:16.291931 8 log.go:172] (0xc0017bb3f0) Data frame received for 3 I0425 00:32:16.291957 8 log.go:172] (0xc0016fcc80) (3) Data frame handling I0425 00:32:16.291978 8 log.go:172] (0xc0016fcc80) (3) Data frame sent I0425 00:32:16.291993 8 log.go:172] (0xc0017bb3f0) Data frame received for 3 I0425 00:32:16.292004 8 log.go:172] (0xc0016fcc80) (3) Data frame handling I0425 00:32:16.292062 8 log.go:172] (0xc0017bb3f0) Data frame received for 5 I0425 00:32:16.292133 8 log.go:172] (0xc0016fcdc0) (5) Data frame handling I0425 00:32:16.293655 8 log.go:172] (0xc0017bb3f0) Data frame received for 1 I0425 00:32:16.293691 8 log.go:172] (0xc001de3040) (1) Data frame handling I0425 00:32:16.293705 8 log.go:172] (0xc001de3040) (1) Data frame sent I0425 00:32:16.293728 8 log.go:172] (0xc0017bb3f0) (0xc001de3040) Stream removed, broadcasting: 1 I0425 00:32:16.293752 8 log.go:172] (0xc0017bb3f0) Go away received I0425 00:32:16.293931 8 log.go:172] (0xc0017bb3f0) (0xc001de3040) Stream removed, broadcasting: 1 I0425 00:32:16.293956 8 log.go:172] (0xc0017bb3f0) (0xc0016fcc80) Stream removed, broadcasting: 3 I0425 00:32:16.293974 8 log.go:172] (0xc0017bb3f0) (0xc0016fcdc0) Stream removed, broadcasting: 5 Apr 25 00:32:16.293: INFO: Exec stderr: "" Apr 25 00:32:16.294: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:16.294: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:16.326450 8 log.go:172] (0xc002dc36b0) (0xc0016fd540) Create stream I0425 00:32:16.326486 8 log.go:172] (0xc002dc36b0) (0xc0016fd540) Stream added, broadcasting: 1 I0425 00:32:16.328614 8 log.go:172] (0xc002dc36b0) Reply frame received for 1 I0425 00:32:16.328659 8 log.go:172] (0xc002dc36b0) (0xc0022be1e0) Create stream I0425 00:32:16.328683 8 log.go:172] (0xc002dc36b0) (0xc0022be1e0) Stream added, broadcasting: 3 I0425 00:32:16.329927 8 log.go:172] (0xc002dc36b0) Reply frame received for 3 I0425 00:32:16.329970 8 log.go:172] (0xc002dc36b0) (0xc0022be320) Create stream I0425 00:32:16.329994 8 log.go:172] (0xc002dc36b0) (0xc0022be320) Stream added, broadcasting: 5 I0425 00:32:16.330932 8 log.go:172] (0xc002dc36b0) Reply frame received for 5 I0425 00:32:16.405954 8 log.go:172] (0xc002dc36b0) Data frame received for 3 I0425 00:32:16.405992 8 log.go:172] (0xc0022be1e0) (3) Data frame handling I0425 00:32:16.406000 8 log.go:172] (0xc0022be1e0) (3) Data frame sent I0425 00:32:16.406006 8 log.go:172] (0xc002dc36b0) Data frame received for 3 I0425 00:32:16.406016 8 log.go:172] (0xc0022be1e0) (3) Data frame handling I0425 00:32:16.406068 8 log.go:172] (0xc002dc36b0) Data frame received for 5 I0425 00:32:16.406097 8 log.go:172] (0xc0022be320) (5) Data frame handling I0425 00:32:16.407222 8 log.go:172] (0xc002dc36b0) Data frame received for 1 I0425 00:32:16.407250 8 log.go:172] (0xc0016fd540) (1) Data frame handling I0425 00:32:16.407263 8 log.go:172] (0xc0016fd540) (1) Data frame sent I0425 00:32:16.407279 8 log.go:172] (0xc002dc36b0) (0xc0016fd540) Stream removed, broadcasting: 1 I0425 00:32:16.407300 8 log.go:172] (0xc002dc36b0) Go away received I0425 00:32:16.407378 8 log.go:172] (0xc002dc36b0) (0xc0016fd540) Stream removed, broadcasting: 1 I0425 00:32:16.407394 8 log.go:172] (0xc002dc36b0) (0xc0022be1e0) Stream removed, broadcasting: 3 I0425 00:32:16.407406 8 log.go:172] (0xc002dc36b0) (0xc0022be320) Stream removed, broadcasting: 5 Apr 25 00:32:16.407: INFO: Exec stderr: "" Apr 25 00:32:16.407: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:16.407: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:16.434919 8 log.go:172] (0xc002dc3ce0) (0xc0016fd9a0) Create stream I0425 00:32:16.434943 8 log.go:172] (0xc002dc3ce0) (0xc0016fd9a0) Stream added, broadcasting: 1 I0425 00:32:16.436627 8 log.go:172] (0xc002dc3ce0) Reply frame received for 1 I0425 00:32:16.436653 8 log.go:172] (0xc002dc3ce0) (0xc0016fda40) Create stream I0425 00:32:16.436668 8 log.go:172] (0xc002dc3ce0) (0xc0016fda40) Stream added, broadcasting: 3 I0425 00:32:16.437747 8 log.go:172] (0xc002dc3ce0) Reply frame received for 3 I0425 00:32:16.437778 8 log.go:172] (0xc002dc3ce0) (0xc0022be460) Create stream I0425 00:32:16.437789 8 log.go:172] (0xc002dc3ce0) (0xc0022be460) Stream added, broadcasting: 5 I0425 00:32:16.438826 8 log.go:172] (0xc002dc3ce0) Reply frame received for 5 I0425 00:32:16.508946 8 log.go:172] (0xc002dc3ce0) Data frame received for 5 I0425 00:32:16.508988 8 log.go:172] (0xc0022be460) (5) Data frame handling I0425 00:32:16.509044 8 log.go:172] (0xc002dc3ce0) Data frame received for 3 I0425 00:32:16.509094 8 log.go:172] (0xc0016fda40) (3) Data frame handling I0425 00:32:16.509105 8 log.go:172] (0xc0016fda40) (3) Data frame sent I0425 00:32:16.509310 8 log.go:172] (0xc002dc3ce0) Data frame received for 3 I0425 00:32:16.509328 8 log.go:172] (0xc0016fda40) (3) Data frame handling I0425 00:32:16.510808 8 log.go:172] (0xc002dc3ce0) Data frame received for 1 I0425 00:32:16.510826 8 log.go:172] (0xc0016fd9a0) (1) Data frame handling I0425 00:32:16.510836 8 log.go:172] (0xc0016fd9a0) (1) Data frame sent I0425 00:32:16.510846 8 log.go:172] (0xc002dc3ce0) (0xc0016fd9a0) Stream removed, broadcasting: 1 I0425 00:32:16.510896 8 log.go:172] (0xc002dc3ce0) Go away received I0425 00:32:16.510940 8 log.go:172] (0xc002dc3ce0) (0xc0016fd9a0) Stream removed, broadcasting: 1 I0425 00:32:16.510955 8 log.go:172] (0xc002dc3ce0) (0xc0016fda40) Stream removed, broadcasting: 3 I0425 00:32:16.510965 8 log.go:172] (0xc002dc3ce0) (0xc0022be460) Stream removed, broadcasting: 5 Apr 25 00:32:16.510: INFO: Exec stderr: "" Apr 25 00:32:16.511: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7219 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 25 00:32:16.511: INFO: >>> kubeConfig: /root/.kube/config I0425 00:32:16.540699 8 log.go:172] (0xc002db42c0) (0xc0016fdcc0) Create stream I0425 00:32:16.540728 8 log.go:172] (0xc002db42c0) (0xc0016fdcc0) Stream added, broadcasting: 1 I0425 00:32:16.542690 8 log.go:172] (0xc002db42c0) Reply frame received for 1 I0425 00:32:16.542744 8 log.go:172] (0xc002db42c0) (0xc0017330e0) Create stream I0425 00:32:16.542761 8 log.go:172] (0xc002db42c0) (0xc0017330e0) Stream added, broadcasting: 3 I0425 00:32:16.543634 8 log.go:172] (0xc002db42c0) Reply frame received for 3 I0425 00:32:16.543683 8 log.go:172] (0xc002db42c0) (0xc001733400) Create stream I0425 00:32:16.543704 8 log.go:172] (0xc002db42c0) (0xc001733400) Stream added, broadcasting: 5 I0425 00:32:16.544664 8 log.go:172] (0xc002db42c0) Reply frame received for 5 I0425 00:32:16.611153 8 log.go:172] (0xc002db42c0) Data frame received for 5 I0425 00:32:16.611187 8 log.go:172] (0xc001733400) (5) Data frame handling I0425 00:32:16.611214 8 log.go:172] (0xc002db42c0) Data frame received for 3 I0425 00:32:16.611243 8 log.go:172] (0xc0017330e0) (3) Data frame handling I0425 00:32:16.611272 8 log.go:172] (0xc0017330e0) (3) Data frame sent I0425 00:32:16.611292 8 log.go:172] (0xc002db42c0) Data frame received for 3 I0425 00:32:16.611300 8 log.go:172] (0xc0017330e0) (3) Data frame handling I0425 00:32:16.612707 8 log.go:172] (0xc002db42c0) Data frame received for 1 I0425 00:32:16.612731 8 log.go:172] (0xc0016fdcc0) (1) Data frame handling I0425 00:32:16.613232 8 log.go:172] (0xc0016fdcc0) (1) Data frame sent I0425 00:32:16.613262 8 log.go:172] (0xc002db42c0) (0xc0016fdcc0) Stream removed, broadcasting: 1 I0425 00:32:16.613290 8 log.go:172] (0xc002db42c0) Go away received I0425 00:32:16.613365 8 log.go:172] (0xc002db42c0) (0xc0016fdcc0) Stream removed, broadcasting: 1 I0425 00:32:16.613378 8 log.go:172] (0xc002db42c0) (0xc0017330e0) Stream removed, broadcasting: 3 I0425 00:32:16.613385 8 log.go:172] (0xc002db42c0) (0xc001733400) Stream removed, broadcasting: 5 Apr 25 00:32:16.613: INFO: Exec stderr: "" [AfterEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:32:16.613: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-kubelet-etc-hosts-7219" for this suite. • [SLOW TEST:13.608 seconds] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":185,"skipped":3349,"failed":0} SSS ------------------------------ [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:32:16.620: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name projected-configmap-test-volume-fd285868-5fc3-4882-9892-a75bafb395a1 STEP: Creating a pod to test consume configMaps Apr 25 00:32:16.700: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b" in namespace "projected-6554" to be "Succeeded or Failed" Apr 25 00:32:16.725: INFO: Pod "pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b": Phase="Pending", Reason="", readiness=false. Elapsed: 24.737969ms Apr 25 00:32:18.729: INFO: Pod "pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028515496s Apr 25 00:32:20.733: INFO: Pod "pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.032621596s STEP: Saw pod success Apr 25 00:32:20.733: INFO: Pod "pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b" satisfied condition "Succeeded or Failed" Apr 25 00:32:20.735: INFO: Trying to get logs from node latest-worker pod pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b container projected-configmap-volume-test: STEP: delete the pod Apr 25 00:32:20.766: INFO: Waiting for pod pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b to disappear Apr 25 00:32:20.785: INFO: Pod pod-projected-configmaps-6e798700-0c99-4171-a8d7-0a3d53367a3b no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:32:20.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6554" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":275,"completed":186,"skipped":3352,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:32:20.797: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:32:21.459: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:32:24.220: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:32:26.225: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371541, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:32:29.249: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource with pruning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:32:29.264: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-4669-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource that should be mutated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:32:30.373: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-8672" for this suite. STEP: Destroying namespace "webhook-8672-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:9.702 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource with pruning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","total":275,"completed":187,"skipped":3430,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:32:30.500: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name projected-configmap-test-volume-aa2cab36-aed9-480d-a61a-61e889dac88a STEP: Creating a pod to test consume configMaps Apr 25 00:32:30.648: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659" in namespace "projected-9662" to be "Succeeded or Failed" Apr 25 00:32:30.834: INFO: Pod "pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659": Phase="Pending", Reason="", readiness=false. Elapsed: 186.052644ms Apr 25 00:32:32.839: INFO: Pod "pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659": Phase="Pending", Reason="", readiness=false. Elapsed: 2.190613925s Apr 25 00:32:34.843: INFO: Pod "pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.194597035s STEP: Saw pod success Apr 25 00:32:34.843: INFO: Pod "pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659" satisfied condition "Succeeded or Failed" Apr 25 00:32:34.845: INFO: Trying to get logs from node latest-worker2 pod pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659 container projected-configmap-volume-test: STEP: delete the pod Apr 25 00:32:34.882: INFO: Waiting for pod pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659 to disappear Apr 25 00:32:34.897: INFO: Pod pod-projected-configmaps-3f068f89-5526-4798-a570-787bcb908659 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:32:34.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9662" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":275,"completed":188,"skipped":3497,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-apps] Job should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:32:34.906: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: delete a job STEP: deleting Job.batch foo in namespace job-1207, will wait for the garbage collector to delete the pods Apr 25 00:32:39.042: INFO: Deleting Job.batch foo took: 6.165191ms Apr 25 00:32:39.342: INFO: Terminating Job.batch foo pods took: 300.287845ms STEP: Ensuring job was deleted [AfterEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:33:23.046: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-1207" for this suite. • [SLOW TEST:48.150 seconds] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","total":275,"completed":189,"skipped":3511,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:33:23.057: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:33:27.176: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-9681" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":190,"skipped":3528,"failed":0} SSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:33:27.184: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:33:27.251: INFO: Creating simple daemon set daemon-set STEP: Check that daemon pods launch on every node of the cluster. Apr 25 00:33:27.258: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:27.272: INFO: Number of nodes with available pods: 0 Apr 25 00:33:27.272: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:33:28.302: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:28.307: INFO: Number of nodes with available pods: 0 Apr 25 00:33:28.307: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:33:29.422: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:29.559: INFO: Number of nodes with available pods: 0 Apr 25 00:33:29.559: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:33:30.277: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:30.280: INFO: Number of nodes with available pods: 0 Apr 25 00:33:30.280: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:33:31.284: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:31.287: INFO: Number of nodes with available pods: 2 Apr 25 00:33:31.287: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Update daemon pods image. STEP: Check that daemon pods images are updated. Apr 25 00:33:31.312: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:31.312: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:31.335: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:32.339: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:32.339: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:32.343: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:33.349: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:33.349: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:33.354: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:34.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:34.340: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:34.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:35.339: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:35.339: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:35.339: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:35.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:36.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:36.340: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:36.340: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:36.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:37.339: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:37.339: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:37.340: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:37.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:38.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:38.340: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:38.341: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:38.345: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:39.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:39.340: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:39.340: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:39.368: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:40.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:40.340: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:40.340: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:40.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:41.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:41.340: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:41.340: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:41.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:42.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:42.340: INFO: Wrong image for pod: daemon-set-t2r7x. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:42.340: INFO: Pod daemon-set-t2r7x is not available Apr 25 00:33:42.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:43.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:43.340: INFO: Pod daemon-set-tg2wc is not available Apr 25 00:33:43.347: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:44.339: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:44.339: INFO: Pod daemon-set-tg2wc is not available Apr 25 00:33:44.342: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:45.344: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:45.344: INFO: Pod daemon-set-tg2wc is not available Apr 25 00:33:45.348: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:46.339: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:46.343: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:47.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:47.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:48.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:48.341: INFO: Pod daemon-set-fgdj5 is not available Apr 25 00:33:48.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:49.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:49.340: INFO: Pod daemon-set-fgdj5 is not available Apr 25 00:33:49.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:50.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:50.340: INFO: Pod daemon-set-fgdj5 is not available Apr 25 00:33:50.343: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:51.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:51.340: INFO: Pod daemon-set-fgdj5 is not available Apr 25 00:33:51.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:52.340: INFO: Wrong image for pod: daemon-set-fgdj5. Expected: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12, got: docker.io/library/httpd:2.4.38-alpine. Apr 25 00:33:52.340: INFO: Pod daemon-set-fgdj5 is not available Apr 25 00:33:52.343: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:53.340: INFO: Pod daemon-set-hw9dt is not available Apr 25 00:33:53.344: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node STEP: Check that daemon pods are still running on every node of the cluster. Apr 25 00:33:53.349: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:53.352: INFO: Number of nodes with available pods: 1 Apr 25 00:33:53.352: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:33:54.434: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:54.438: INFO: Number of nodes with available pods: 1 Apr 25 00:33:54.438: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:33:55.358: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:55.362: INFO: Number of nodes with available pods: 1 Apr 25 00:33:55.362: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:33:56.358: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:33:56.361: INFO: Number of nodes with available pods: 2 Apr 25 00:33:56.361: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-138, will wait for the garbage collector to delete the pods Apr 25 00:33:56.433: INFO: Deleting DaemonSet.extensions daemon-set took: 6.376698ms Apr 25 00:33:56.733: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.253654ms Apr 25 00:34:03.037: INFO: Number of nodes with available pods: 0 Apr 25 00:34:03.037: INFO: Number of running nodes: 0, number of available pods: 0 Apr 25 00:34:03.039: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-138/daemonsets","resourceVersion":"10796746"},"items":null} Apr 25 00:34:03.042: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-138/pods","resourceVersion":"10796746"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:34:03.050: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-138" for this suite. • [SLOW TEST:35.872 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","total":275,"completed":191,"skipped":3534,"failed":0} [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:34:03.056: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a replication controller. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ReplicationController STEP: Ensuring resource quota status captures replication controller creation STEP: Deleting a ReplicationController STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:34:14.193: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-3830" for this suite. • [SLOW TEST:11.151 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a replication controller. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","total":275,"completed":192,"skipped":3534,"failed":0} SSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:34:14.207: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Apr 25 00:34:24.372: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 25 00:34:24.379: INFO: Pod pod-with-poststart-http-hook still exists Apr 25 00:34:26.379: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 25 00:34:26.384: INFO: Pod pod-with-poststart-http-hook still exists Apr 25 00:34:28.379: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 25 00:34:28.385: INFO: Pod pod-with-poststart-http-hook still exists Apr 25 00:34:30.379: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 25 00:34:30.384: INFO: Pod pod-with-poststart-http-hook still exists Apr 25 00:34:32.379: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 25 00:34:32.383: INFO: Pod pod-with-poststart-http-hook still exists Apr 25 00:34:34.379: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 25 00:34:34.383: INFO: Pod pod-with-poststart-http-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:34:34.383: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-9416" for this suite. • [SLOW TEST:20.185 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","total":275,"completed":193,"skipped":3544,"failed":0} SSSS ------------------------------ [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:34:34.392: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:34:34.509: INFO: Create a RollingUpdate DaemonSet Apr 25 00:34:34.513: INFO: Check that daemon pods launch on every node of the cluster Apr 25 00:34:34.518: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:34.523: INFO: Number of nodes with available pods: 0 Apr 25 00:34:34.523: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:34:35.528: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:35.530: INFO: Number of nodes with available pods: 0 Apr 25 00:34:35.530: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:34:36.528: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:36.531: INFO: Number of nodes with available pods: 0 Apr 25 00:34:36.531: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:34:38.237: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:38.795: INFO: Number of nodes with available pods: 1 Apr 25 00:34:38.795: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:34:39.527: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:39.530: INFO: Number of nodes with available pods: 1 Apr 25 00:34:39.530: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:34:40.528: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:40.532: INFO: Number of nodes with available pods: 2 Apr 25 00:34:40.532: INFO: Number of running nodes: 2, number of available pods: 2 Apr 25 00:34:40.532: INFO: Update the DaemonSet to trigger a rollout Apr 25 00:34:40.540: INFO: Updating DaemonSet daemon-set Apr 25 00:34:53.560: INFO: Roll back the DaemonSet before rollout is complete Apr 25 00:34:53.596: INFO: Updating DaemonSet daemon-set Apr 25 00:34:53.596: INFO: Make sure DaemonSet rollback is complete Apr 25 00:34:53.614: INFO: Wrong image for pod: daemon-set-grhq5. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Apr 25 00:34:53.614: INFO: Pod daemon-set-grhq5 is not available Apr 25 00:34:53.634: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:54.639: INFO: Wrong image for pod: daemon-set-grhq5. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Apr 25 00:34:54.639: INFO: Pod daemon-set-grhq5 is not available Apr 25 00:34:54.642: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:55.638: INFO: Wrong image for pod: daemon-set-grhq5. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Apr 25 00:34:55.638: INFO: Pod daemon-set-grhq5 is not available Apr 25 00:34:55.642: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:34:56.704: INFO: Pod daemon-set-xpww5 is not available Apr 25 00:34:56.708: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1257, will wait for the garbage collector to delete the pods Apr 25 00:34:56.770: INFO: Deleting DaemonSet.extensions daemon-set took: 5.481066ms Apr 25 00:34:57.071: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.264238ms Apr 25 00:34:59.773: INFO: Number of nodes with available pods: 0 Apr 25 00:34:59.773: INFO: Number of running nodes: 0, number of available pods: 0 Apr 25 00:34:59.776: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-1257/daemonsets","resourceVersion":"10797086"},"items":null} Apr 25 00:34:59.778: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-1257/pods","resourceVersion":"10797086"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:34:59.787: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-1257" for this suite. • [SLOW TEST:25.403 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","total":275,"completed":194,"skipped":3548,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:34:59.795: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:34:59.860: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e" in namespace "downward-api-52" to be "Succeeded or Failed" Apr 25 00:34:59.876: INFO: Pod "downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e": Phase="Pending", Reason="", readiness=false. Elapsed: 15.828244ms Apr 25 00:35:01.880: INFO: Pod "downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01968976s Apr 25 00:35:03.883: INFO: Pod "downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023552783s STEP: Saw pod success Apr 25 00:35:03.883: INFO: Pod "downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e" satisfied condition "Succeeded or Failed" Apr 25 00:35:03.887: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e container client-container: STEP: delete the pod Apr 25 00:35:03.915: INFO: Waiting for pod downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e to disappear Apr 25 00:35:03.925: INFO: Pod downwardapi-volume-3824402d-1616-45ee-bf62-eb753306135e no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:35:03.926: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-52" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","total":275,"completed":195,"skipped":3565,"failed":0} SSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:35:03.932: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:84 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:99 STEP: Creating service test in namespace statefulset-4757 [It] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a new StatefulSet Apr 25 00:35:04.033: INFO: Found 0 stateful pods, waiting for 3 Apr 25 00:35:14.040: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:35:14.040: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:35:14.040: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=false Apr 25 00:35:24.038: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:35:24.038: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:35:24.038: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Updating stateful set template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine Apr 25 00:35:24.066: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Not applying an update when the partition is greater than the number of replicas STEP: Performing a canary update Apr 25 00:35:34.102: INFO: Updating stateful set ss2 Apr 25 00:35:34.160: INFO: Waiting for Pod statefulset-4757/ss2-2 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 STEP: Restoring Pods to the correct revision when they are deleted Apr 25 00:35:44.616: INFO: Found 2 stateful pods, waiting for 3 Apr 25 00:35:54.621: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:35:54.621: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 25 00:35:54.621: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Performing a phased rolling update Apr 25 00:35:54.644: INFO: Updating stateful set ss2 Apr 25 00:35:54.664: INFO: Waiting for Pod statefulset-4757/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Apr 25 00:36:05.567: INFO: Updating stateful set ss2 Apr 25 00:36:05.652: INFO: Waiting for StatefulSet statefulset-4757/ss2 to complete update Apr 25 00:36:05.652: INFO: Waiting for Pod statefulset-4757/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:110 Apr 25 00:36:15.659: INFO: Deleting all statefulset in ns statefulset-4757 Apr 25 00:36:15.661: INFO: Scaling statefulset ss2 to 0 Apr 25 00:36:45.686: INFO: Waiting for statefulset status.replicas updated to 0 Apr 25 00:36:45.689: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:36:45.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-4757" for this suite. • [SLOW TEST:101.796 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","total":275,"completed":196,"skipped":3572,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:36:45.729: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:74 [It] deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:36:45.775: INFO: Creating deployment "webserver-deployment" Apr 25 00:36:45.789: INFO: Waiting for observed generation 1 Apr 25 00:36:48.438: INFO: Waiting for all required pods to come up Apr 25 00:36:48.689: INFO: Pod name httpd: Found 10 pods out of 10 STEP: ensuring each pod is running Apr 25 00:36:56.971: INFO: Waiting for deployment "webserver-deployment" to complete Apr 25 00:36:56.975: INFO: Updating deployment "webserver-deployment" with a non-existent image Apr 25 00:36:56.993: INFO: Updating deployment webserver-deployment Apr 25 00:36:56.993: INFO: Waiting for observed generation 2 Apr 25 00:36:59.696: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 Apr 25 00:36:59.944: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 Apr 25 00:37:00.408: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas Apr 25 00:37:00.618: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 Apr 25 00:37:00.618: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 Apr 25 00:37:00.713: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas Apr 25 00:37:00.933: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas Apr 25 00:37:00.933: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 Apr 25 00:37:01.030: INFO: Updating deployment webserver-deployment Apr 25 00:37:01.030: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas Apr 25 00:37:01.283: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 Apr 25 00:37:04.282: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:68 Apr 25 00:37:04.688: INFO: Deployment "webserver-deployment": &Deployment{ObjectMeta:{webserver-deployment deployment-1513 /apis/apps/v1/namespaces/deployment-1513/deployments/webserver-deployment 2c5f9a51-03c7-4c7c-bf79-aeef604f42e7 10797985 3 2020-04-25 00:36:45 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] []},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0035897d8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-04-25 00:37:01 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-c7997dcc8" is progressing.,LastUpdateTime:2020-04-25 00:37:01 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} Apr 25 00:37:04.752: INFO: New ReplicaSet "webserver-deployment-c7997dcc8" of Deployment "webserver-deployment": &ReplicaSet{ObjectMeta:{webserver-deployment-c7997dcc8 deployment-1513 /apis/apps/v1/namespaces/deployment-1513/replicasets/webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 10797982 3 2020-04-25 00:36:56 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment 2c5f9a51-03c7-4c7c-bf79-aeef604f42e7 0xc003589d17 0xc003589d18}] [] []},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: c7997dcc8,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc003589d88 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:37:04.752: INFO: All old ReplicaSets of Deployment "webserver-deployment": Apr 25 00:37:04.752: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-595b5b9587 deployment-1513 /apis/apps/v1/namespaces/deployment-1513/replicasets/webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 10797974 3 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment 2c5f9a51-03c7-4c7c-bf79-aeef604f42e7 0xc003589c57 0xc003589c58}] [] []},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 595b5b9587,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc003589cb8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:37:04.857: INFO: Pod "webserver-deployment-595b5b9587-4jh2p" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-4jh2p webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-4jh2p 9f6df5b5-cac3-4441-8e2f-6a64682149e9 10798006 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f4307 0xc0033f4308}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.857: INFO: Pod "webserver-deployment-595b5b9587-6fmwk" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-6fmwk webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-6fmwk f1b09f7f-a166-4118-a9af-97e8bc5c49f2 10797993 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f4467 0xc0033f4468}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.857: INFO: Pod "webserver-deployment-595b5b9587-85xd4" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-85xd4 webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-85xd4 8c014c33-fee9-42ed-9e45-41846b3c893e 10797784 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f4647 0xc0033f4648}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.213,StartTime:2020-04-25 00:36:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://82f6eb98338d5b32be06752bd2554db2eb06502c9da05048696dbfd140624596,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.213,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.857: INFO: Pod "webserver-deployment-595b5b9587-8jmw5" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-8jmw5 webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-8jmw5 3427bec7-b3f3-42eb-8fa9-fdf4552dfe6a 10797960 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f4967 0xc0033f4968}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.858: INFO: Pod "webserver-deployment-595b5b9587-9pq2m" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-9pq2m webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-9pq2m df259c71-d770-4d02-96f7-acfcfd677ea6 10797794 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f4b17 0xc0033f4b18}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:10.244.1.249,StartTime:2020-04-25 00:36:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://0abecbd5f8d55b38b177059cc5cd80dead22f9279608a4eb2bd917201edad4d6,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.249,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.858: INFO: Pod "webserver-deployment-595b5b9587-cnzsb" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-cnzsb webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-cnzsb a3204c25-ec62-462b-849d-80a42cff660b 10797758 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f4dc7 0xc0033f4dc8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.210,StartTime:2020-04-25 00:36:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:52 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://295f1776aa10741bd1350a44dbb0171939e0f5c7df8a1b7b3469a806889b1691,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.210,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.858: INFO: Pod "webserver-deployment-595b5b9587-g5dqt" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-g5dqt webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-g5dqt 60af77a7-5188-428f-988b-7970d168be29 10797962 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f4fb7 0xc0033f4fb8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.858: INFO: Pod "webserver-deployment-595b5b9587-h9d7j" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-h9d7j webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-h9d7j 4a5700eb-97a8-4a3a-aa9e-1ff616e57fdb 10797745 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f50e7 0xc0033f50e8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:51 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:51 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.209,StartTime:2020-04-25 00:36:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:49 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://0f2f4c98fb944bf05bfc34676efb842ec597bf42db02644a837c5313fdb14323,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.209,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.858: INFO: Pod "webserver-deployment-595b5b9587-hcswg" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-hcswg webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-hcswg e50ad14d-eb73-4094-bad0-4380988df49f 10797957 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5267 0xc0033f5268}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.859: INFO: Pod "webserver-deployment-595b5b9587-jt944" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-jt944 webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-jt944 afc4fd0f-896e-421a-ae83-43b5afff3424 10798024 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5387 0xc0033f5388}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:,StartTime:2020-04-25 00:37:03 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.859: INFO: Pod "webserver-deployment-595b5b9587-jtxx7" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-jtxx7 webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-jtxx7 bfc5ab34-d437-41df-9756-a5e23cd55437 10797770 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f54e7 0xc0033f54e8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:53 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:53 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.211,StartTime:2020-04-25 00:36:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:52 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://c1ab0d412cb2b948c7d50d83d65a48794c17d44a65a1e719adb26a99d6c69a57,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.211,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.859: INFO: Pod "webserver-deployment-595b5b9587-l2kbx" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-l2kbx webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-l2kbx 959de633-9b05-4a71-807d-627ef22a434c 10797999 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5667 0xc0033f5668}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.859: INFO: Pod "webserver-deployment-595b5b9587-ngc4w" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-ngc4w webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-ngc4w 5ea9ed0e-b9a7-4272-b122-61d355be2174 10797779 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f57c7 0xc0033f57c8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:53 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:53 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:10.244.1.246,StartTime:2020-04-25 00:36:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:53 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://710c309899171250628bfc3b741568a69363464f5cc1aad6390399300239d20f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.246,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.859: INFO: Pod "webserver-deployment-595b5b9587-nj48d" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-nj48d webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-nj48d 9472e0db-777a-4d7b-be27-4ad12d8b6ff9 10797944 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5947 0xc0033f5948}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.860: INFO: Pod "webserver-deployment-595b5b9587-nmgp6" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-nmgp6 webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-nmgp6 676ee37e-8083-4390-ae1c-0fef7c4078da 10797995 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5aa7 0xc0033f5aa8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.860: INFO: Pod "webserver-deployment-595b5b9587-pth49" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-pth49 webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-pth49 8b520e4b-332a-4999-b11a-b3ddc0e6e4d4 10797963 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5c17 0xc0033f5c18}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.860: INFO: Pod "webserver-deployment-595b5b9587-qsrrd" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-qsrrd webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-qsrrd c611c8b4-93f8-419d-8688-9ef24b03e4db 10797988 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5d37 0xc0033f5d38}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.860: INFO: Pod "webserver-deployment-595b5b9587-xczxm" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-xczxm webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-xczxm 4c6590a8-fd55-4a0c-82bd-3948a34de6ea 10797810 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc0033f5e97 0xc0033f5e98}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.212,StartTime:2020-04-25 00:36:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:55 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://f0ec33cbbcefa294f6022d817f4ee593a032b16384e06ec5329bcddfd2516b0f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.212,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.860: INFO: Pod "webserver-deployment-595b5b9587-zvft2" is not available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-zvft2 webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-zvft2 3f1966d1-c53e-4abf-9fd7-9faee548c2b7 10797938 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc00360a017 0xc00360a018}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.860: INFO: Pod "webserver-deployment-595b5b9587-zzqgs" is available: &Pod{ObjectMeta:{webserver-deployment-595b5b9587-zzqgs webserver-deployment-595b5b9587- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-595b5b9587-zzqgs 1ae4d39e-b8f6-42e9-9aa8-002cfffbbb52 10797801 0 2020-04-25 00:36:45 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 60713010-1520-4bff-bca7-beb652bb6247 0xc00360a157 0xc00360a158}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:10.244.1.247,StartTime:2020-04-25 00:36:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:36:53 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://75bd98c449a2779e889270d7e8f292b1e2460ff07a252e3df3114b8c4b15d0c0,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.247,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.860: INFO: Pod "webserver-deployment-c7997dcc8-4dclb" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-4dclb webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-4dclb 0075c289-e443-4415-a6c8-02a6785cfd6c 10798021 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360a2d7 0xc00360a2d8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:04 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:04 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:37:04 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.861: INFO: Pod "webserver-deployment-c7997dcc8-5dz6f" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-5dz6f webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-5dz6f ad563065-e391-4aa1-ba56-ea4fd1744d7d 10797930 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360a457 0xc00360a458}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.861: INFO: Pod "webserver-deployment-c7997dcc8-6sjkq" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-6sjkq webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-6sjkq 124f11ce-c91f-410b-8b17-5039817e4173 10797863 0 2020-04-25 00:36:57 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360a587 0xc00360a588}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:36:57 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.861: INFO: Pod "webserver-deployment-c7997dcc8-88pwr" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-88pwr webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-88pwr 82547853-32d2-4aad-a7f6-6b60f89ad355 10797942 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360a707 0xc00360a708}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.861: INFO: Pod "webserver-deployment-c7997dcc8-8jbf7" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-8jbf7 webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-8jbf7 a58143bc-6794-4bb0-a4f0-09e457deeae1 10797887 0 2020-04-25 00:36:59 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360a837 0xc00360a838}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:,StartTime:2020-04-25 00:36:59 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.861: INFO: Pod "webserver-deployment-c7997dcc8-ftzxt" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-ftzxt webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-ftzxt 39b1073c-7acc-41bc-99e7-e9773bb755af 10797890 0 2020-04-25 00:36:59 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360a9c7 0xc00360a9c8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:59 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:36:59 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.862: INFO: Pod "webserver-deployment-c7997dcc8-jrjvw" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-jrjvw webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-jrjvw 3bf45af1-fbdb-4803-b433-59a3271eeccd 10798014 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360ab57 0xc00360ab58}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:02 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:02 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:,StartTime:2020-04-25 00:37:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.862: INFO: Pod "webserver-deployment-c7997dcc8-lwt5t" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-lwt5t webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-lwt5t 620f9450-ac6e-4351-8bf8-5baa424c5814 10797979 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360ace7 0xc00360ace8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.862: INFO: Pod "webserver-deployment-c7997dcc8-lxrpw" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-lxrpw webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-lxrpw 31174561-a39f-44b6-8cf8-579e329cd466 10797855 0 2020-04-25 00:36:57 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360ae67 0xc00360ae68}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:36:57 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.862: INFO: Pod "webserver-deployment-c7997dcc8-nd74g" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-nd74g webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-nd74g 3c04adcc-9e09-4adf-a918-e595db71228a 10797966 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360afe7 0xc00360afe8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.862: INFO: Pod "webserver-deployment-c7997dcc8-p5grx" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-p5grx webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-p5grx 7ced284f-f278-4dce-9cb0-17a475cd9d65 10798011 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360b117 0xc00360b118}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:37:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.862: INFO: Pod "webserver-deployment-c7997dcc8-ppn7n" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-ppn7n webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-ppn7n 9e07c021-3e6e-4681-b50b-118a096a830e 10797991 0 2020-04-25 00:36:57 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360b2a7 0xc00360b2a8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:36:57 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.214,StartTime:2020-04-25 00:36:57 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = failed to pull and unpack image "docker.io/library/webserver:404": failed to resolve reference "docker.io/library/webserver:404": pull access denied, repository does not exist or may require authorization: server message: insufficient_scope: authorization failed,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.214,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Apr 25 00:37:04.862: INFO: Pod "webserver-deployment-c7997dcc8-rcfx4" is not available: &Pod{ObjectMeta:{webserver-deployment-c7997dcc8-rcfx4 webserver-deployment-c7997dcc8- deployment-1513 /api/v1/namespaces/deployment-1513/pods/webserver-deployment-c7997dcc8-rcfx4 0e6562f4-7e82-4bd8-8421-7b36180e5923 10797946 0 2020-04-25 00:37:01 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 1883550d-9d6a-43b6-ae6e-523ae65700fb 0xc00360b457 0xc00360b458}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-jrm4p,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-jrm4p,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-jrm4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:37:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:37:04.863: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-1513" for this suite. • [SLOW TEST:19.251 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":275,"completed":197,"skipped":3612,"failed":0} SSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:37:04.981: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:37:09.353: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:37:12.553: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371828, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:37:14.826: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371828, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:37:16.811: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371829, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723371828, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:37:19.576: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API STEP: Creating a dummy validating-webhook-configuration object STEP: Deleting the validating-webhook-configuration, which should be possible to remove STEP: Creating a dummy mutating-webhook-configuration object STEP: Deleting the mutating-webhook-configuration, which should be possible to remove [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:37:19.772: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-5708" for this suite. STEP: Destroying namespace "webhook-5708-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:15.039 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","total":275,"completed":198,"skipped":3615,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:37:20.020: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod test-webserver-314fc917-f92c-4aa9-844a-bfaed797801c in namespace container-probe-2904 Apr 25 00:37:26.302: INFO: Started pod test-webserver-314fc917-f92c-4aa9-844a-bfaed797801c in namespace container-probe-2904 STEP: checking the pod's current state and verifying that restartCount is present Apr 25 00:37:26.331: INFO: Initial restart count of pod test-webserver-314fc917-f92c-4aa9-844a-bfaed797801c is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:41:27.256: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-2904" for this suite. • [SLOW TEST:247.284 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":275,"completed":199,"skipped":3642,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:41:27.304: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a watch on configmaps with a certain label STEP: creating a new configmap STEP: modifying the configmap once STEP: changing the label value of the configmap STEP: Expecting to observe a delete notification for the watched object Apr 25 00:41:27.378: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-2208 /api/v1/namespaces/watch-2208/configmaps/e2e-watch-test-label-changed 4f8c4d22-de3d-444c-8bd3-716f513e60f5 10799194 0 2020-04-25 00:41:27 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:41:27.379: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-2208 /api/v1/namespaces/watch-2208/configmaps/e2e-watch-test-label-changed 4f8c4d22-de3d-444c-8bd3-716f513e60f5 10799195 0 2020-04-25 00:41:27 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:41:27.379: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-2208 /api/v1/namespaces/watch-2208/configmaps/e2e-watch-test-label-changed 4f8c4d22-de3d-444c-8bd3-716f513e60f5 10799196 0 2020-04-25 00:41:27 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying the configmap a second time STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements STEP: changing the label value of the configmap back STEP: modifying the configmap a third time STEP: deleting the configmap STEP: Expecting to observe an add notification for the watched object when the label value was restored Apr 25 00:41:37.414: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-2208 /api/v1/namespaces/watch-2208/configmaps/e2e-watch-test-label-changed 4f8c4d22-de3d-444c-8bd3-716f513e60f5 10799236 0 2020-04-25 00:41:27 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:41:37.414: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-2208 /api/v1/namespaces/watch-2208/configmaps/e2e-watch-test-label-changed 4f8c4d22-de3d-444c-8bd3-716f513e60f5 10799237 0 2020-04-25 00:41:27 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:41:37.414: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-2208 /api/v1/namespaces/watch-2208/configmaps/e2e-watch-test-label-changed 4f8c4d22-de3d-444c-8bd3-716f513e60f5 10799238 0 2020-04-25 00:41:27 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:41:37.415: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-2208" for this suite. • [SLOW TEST:10.136 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":275,"completed":200,"skipped":3666,"failed":0} SSS ------------------------------ [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:41:37.441: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-volume-28c3e331-3f3a-4650-b67d-fb03f0e5db45 STEP: Creating a pod to test consume configMaps Apr 25 00:41:37.514: INFO: Waiting up to 5m0s for pod "pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4" in namespace "configmap-9026" to be "Succeeded or Failed" Apr 25 00:41:37.528: INFO: Pod "pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4": Phase="Pending", Reason="", readiness=false. Elapsed: 14.424322ms Apr 25 00:41:39.631: INFO: Pod "pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.117493863s Apr 25 00:41:41.636: INFO: Pod "pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.12169978s STEP: Saw pod success Apr 25 00:41:41.636: INFO: Pod "pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4" satisfied condition "Succeeded or Failed" Apr 25 00:41:41.643: INFO: Trying to get logs from node latest-worker pod pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4 container configmap-volume-test: STEP: delete the pod Apr 25 00:41:41.678: INFO: Waiting for pod pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4 to disappear Apr 25 00:41:41.690: INFO: Pod pod-configmaps-297a103f-c6fb-4da0-afa2-3f099c3ef3c4 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:41:41.690: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-9026" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":275,"completed":201,"skipped":3669,"failed":0} SSSS ------------------------------ [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:41:41.698: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:41:41.765: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137" in namespace "downward-api-870" to be "Succeeded or Failed" Apr 25 00:41:41.774: INFO: Pod "downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137": Phase="Pending", Reason="", readiness=false. Elapsed: 8.651753ms Apr 25 00:41:43.778: INFO: Pod "downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013169114s Apr 25 00:41:45.782: INFO: Pod "downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016546585s STEP: Saw pod success Apr 25 00:41:45.782: INFO: Pod "downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137" satisfied condition "Succeeded or Failed" Apr 25 00:41:45.785: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137 container client-container: STEP: delete the pod Apr 25 00:41:45.806: INFO: Waiting for pod downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137 to disappear Apr 25 00:41:45.810: INFO: Pod downwardapi-volume-7d671191-264b-4e2e-863a-a02cd2220137 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:41:45.810: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-870" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","total":275,"completed":202,"skipped":3673,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:41:45.819: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name projected-configmap-test-volume-map-5dd3800e-308f-482d-8112-503066fdc275 STEP: Creating a pod to test consume configMaps Apr 25 00:41:45.908: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f" in namespace "projected-6250" to be "Succeeded or Failed" Apr 25 00:41:45.918: INFO: Pod "pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f": Phase="Pending", Reason="", readiness=false. Elapsed: 9.397649ms Apr 25 00:41:47.931: INFO: Pod "pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022595408s Apr 25 00:41:49.935: INFO: Pod "pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026827854s STEP: Saw pod success Apr 25 00:41:49.935: INFO: Pod "pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f" satisfied condition "Succeeded or Failed" Apr 25 00:41:49.939: INFO: Trying to get logs from node latest-worker pod pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f container projected-configmap-volume-test: STEP: delete the pod Apr 25 00:41:50.238: INFO: Waiting for pod pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f to disappear Apr 25 00:41:50.250: INFO: Pod pod-projected-configmaps-76f39b26-aa35-4ae4-bfbe-8bd9098fb91f no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:41:50.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6250" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":275,"completed":203,"skipped":3695,"failed":0} SSSSSS ------------------------------ [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:41:50.258: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: starting an echo server on multiple ports STEP: creating replication controller proxy-service-wvcgb in namespace proxy-2545 I0425 00:41:50.676859 8 runners.go:190] Created replication controller with name: proxy-service-wvcgb, namespace: proxy-2545, replica count: 1 I0425 00:41:51.727323 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0425 00:41:52.727546 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0425 00:41:53.727815 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:41:54.728068 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:41:55.728324 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:41:56.728705 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:41:57.729096 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:41:58.729580 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:41:59.729771 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:42:00.730010 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:42:01.730203 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0425 00:42:02.730418 8 runners.go:190] proxy-service-wvcgb Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Apr 25 00:42:02.734: INFO: setup took 12.098508744s, starting test cases STEP: running 16 cases, 20 attempts per case, 320 total attempts Apr 25 00:42:02.739: INFO: (0) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 5.37496ms) Apr 25 00:42:02.740: INFO: (0) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 6.306312ms) Apr 25 00:42:02.744: INFO: (0) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 10.737173ms) Apr 25 00:42:02.744: INFO: (0) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 10.793484ms) Apr 25 00:42:02.745: INFO: (0) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 10.866681ms) Apr 25 00:42:02.746: INFO: (0) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 12.685464ms) Apr 25 00:42:02.746: INFO: (0) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 12.718702ms) Apr 25 00:42:02.747: INFO: (0) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 12.899025ms) Apr 25 00:42:02.747: INFO: (0) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 12.785108ms) Apr 25 00:42:02.747: INFO: (0) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 13.020974ms) Apr 25 00:42:02.748: INFO: (0) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 13.831153ms) Apr 25 00:42:02.749: INFO: (0) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test (200; 2.924425ms) Apr 25 00:42:02.759: INFO: (1) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.248015ms) Apr 25 00:42:02.759: INFO: (1) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.436288ms) Apr 25 00:42:02.759: INFO: (1) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.351782ms) Apr 25 00:42:02.760: INFO: (1) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.033597ms) Apr 25 00:42:02.760: INFO: (1) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 5.068429ms) Apr 25 00:42:02.760: INFO: (1) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.18618ms) Apr 25 00:42:02.760: INFO: (1) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: ... (200; 4.076113ms) Apr 25 00:42:02.765: INFO: (2) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 4.063019ms) Apr 25 00:42:02.765: INFO: (2) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.113475ms) Apr 25 00:42:02.765: INFO: (2) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.084371ms) Apr 25 00:42:02.765: INFO: (2) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test (200; 4.762651ms) Apr 25 00:42:02.766: INFO: (2) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.73065ms) Apr 25 00:42:02.766: INFO: (2) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 4.732771ms) Apr 25 00:42:02.766: INFO: (2) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.103933ms) Apr 25 00:42:02.766: INFO: (2) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 5.06132ms) Apr 25 00:42:02.767: INFO: (2) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.646327ms) Apr 25 00:42:02.767: INFO: (2) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 5.64757ms) Apr 25 00:42:02.767: INFO: (2) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.799127ms) Apr 25 00:42:02.767: INFO: (2) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.737274ms) Apr 25 00:42:02.767: INFO: (2) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 5.791999ms) Apr 25 00:42:02.770: INFO: (3) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 3.205159ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.668028ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 4.616446ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.716541ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.171373ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 5.0564ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.075918ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.196729ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 5.169335ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.21121ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.238507ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 5.317009ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.261516ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 5.269539ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 5.334878ms) Apr 25 00:42:02.772: INFO: (3) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test (200; 3.101464ms) Apr 25 00:42:02.777: INFO: (4) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.299082ms) Apr 25 00:42:02.779: INFO: (4) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 6.454176ms) Apr 25 00:42:02.779: INFO: (4) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 6.43124ms) Apr 25 00:42:02.779: INFO: (4) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 6.508365ms) Apr 25 00:42:02.779: INFO: (4) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 6.667008ms) Apr 25 00:42:02.779: INFO: (4) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 6.690155ms) Apr 25 00:42:02.779: INFO: (4) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test (200; 4.493804ms) Apr 25 00:42:02.785: INFO: (5) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 4.966965ms) Apr 25 00:42:02.785: INFO: (5) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: ... (200; 5.042269ms) Apr 25 00:42:02.785: INFO: (5) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.133163ms) Apr 25 00:42:02.785: INFO: (5) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 5.150026ms) Apr 25 00:42:02.785: INFO: (5) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 5.090968ms) Apr 25 00:42:02.785: INFO: (5) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.165359ms) Apr 25 00:42:02.785: INFO: (5) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.078076ms) Apr 25 00:42:02.786: INFO: (5) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 6.091318ms) Apr 25 00:42:02.786: INFO: (5) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 6.23232ms) Apr 25 00:42:02.786: INFO: (5) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 6.264743ms) Apr 25 00:42:02.786: INFO: (5) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 6.338787ms) Apr 25 00:42:02.786: INFO: (5) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 6.368841ms) Apr 25 00:42:02.786: INFO: (5) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 6.357148ms) Apr 25 00:42:02.791: INFO: (6) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 3.820828ms) Apr 25 00:42:02.791: INFO: (6) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.259321ms) Apr 25 00:42:02.791: INFO: (6) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.506144ms) Apr 25 00:42:02.792: INFO: (6) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 5.785661ms) Apr 25 00:42:02.792: INFO: (6) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 5.704104ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 5.630376ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 4.580003ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.812567ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.995614ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.86003ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.953934ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test (200; 4.961892ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.082291ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 5.215778ms) Apr 25 00:42:02.793: INFO: (6) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.267559ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 3.651878ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 3.634662ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 3.689927ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 3.659495ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 3.694694ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 3.702739ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 3.71609ms) Apr 25 00:42:02.797: INFO: (7) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 3.745053ms) Apr 25 00:42:02.798: INFO: (7) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 3.988532ms) Apr 25 00:42:02.798: INFO: (7) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: ... (200; 3.348551ms) Apr 25 00:42:02.803: INFO: (8) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 3.358464ms) Apr 25 00:42:02.803: INFO: (8) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 3.402923ms) Apr 25 00:42:02.803: INFO: (8) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 3.43513ms) Apr 25 00:42:02.803: INFO: (8) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 3.325674ms) Apr 25 00:42:02.804: INFO: (8) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 4.793504ms) Apr 25 00:42:02.804: INFO: (8) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 4.817488ms) Apr 25 00:42:02.804: INFO: (8) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 4.866171ms) Apr 25 00:42:02.804: INFO: (8) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 4.900428ms) Apr 25 00:42:02.804: INFO: (8) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 4.85365ms) Apr 25 00:42:02.804: INFO: (8) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 4.946626ms) Apr 25 00:42:02.808: INFO: (9) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 3.979056ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 4.263211ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.31011ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 4.344836ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 4.630599ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.708816ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 4.698881ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.694969ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 4.705921ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.74517ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test<... (200; 4.91503ms) Apr 25 00:42:02.809: INFO: (9) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 5.08803ms) Apr 25 00:42:02.810: INFO: (9) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.304673ms) Apr 25 00:42:02.810: INFO: (9) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.229558ms) Apr 25 00:42:02.813: INFO: (10) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 3.461604ms) Apr 25 00:42:02.813: INFO: (10) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 3.788219ms) Apr 25 00:42:02.813: INFO: (10) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 3.828566ms) Apr 25 00:42:02.813: INFO: (10) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: ... (200; 3.828024ms) Apr 25 00:42:02.814: INFO: (10) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 4.355012ms) Apr 25 00:42:02.814: INFO: (10) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 4.359313ms) Apr 25 00:42:02.815: INFO: (10) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 5.236227ms) Apr 25 00:42:02.815: INFO: (10) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.271277ms) Apr 25 00:42:02.815: INFO: (10) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 5.373712ms) Apr 25 00:42:02.815: INFO: (10) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.358562ms) Apr 25 00:42:02.815: INFO: (10) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.422605ms) Apr 25 00:42:02.815: INFO: (10) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 5.499963ms) Apr 25 00:42:02.815: INFO: (10) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 5.414817ms) Apr 25 00:42:02.817: INFO: (11) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test<... (200; 4.334284ms) Apr 25 00:42:02.819: INFO: (11) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.285528ms) Apr 25 00:42:02.819: INFO: (11) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.320592ms) Apr 25 00:42:02.819: INFO: (11) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.306345ms) Apr 25 00:42:02.820: INFO: (11) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.321788ms) Apr 25 00:42:02.820: INFO: (11) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 4.34184ms) Apr 25 00:42:02.820: INFO: (11) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.382653ms) Apr 25 00:42:02.820: INFO: (11) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.357163ms) Apr 25 00:42:02.821: INFO: (11) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 5.38922ms) Apr 25 00:42:02.821: INFO: (11) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.638147ms) Apr 25 00:42:02.821: INFO: (11) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.808931ms) Apr 25 00:42:02.821: INFO: (11) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 5.864081ms) Apr 25 00:42:02.821: INFO: (11) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 5.799822ms) Apr 25 00:42:02.821: INFO: (11) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.914129ms) Apr 25 00:42:02.824: INFO: (12) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 2.827862ms) Apr 25 00:42:02.825: INFO: (12) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 3.574062ms) Apr 25 00:42:02.825: INFO: (12) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 3.826477ms) Apr 25 00:42:02.825: INFO: (12) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 3.934643ms) Apr 25 00:42:02.825: INFO: (12) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.09029ms) Apr 25 00:42:02.825: INFO: (12) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.182397ms) Apr 25 00:42:02.825: INFO: (12) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.099778ms) Apr 25 00:42:02.825: INFO: (12) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 4.336549ms) Apr 25 00:42:02.826: INFO: (12) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test<... (200; 4.511356ms) Apr 25 00:42:02.828: INFO: (13) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 2.317509ms) Apr 25 00:42:02.828: INFO: (13) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 2.417135ms) Apr 25 00:42:02.829: INFO: (13) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 3.530951ms) Apr 25 00:42:02.830: INFO: (13) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test<... (200; 4.229852ms) Apr 25 00:42:02.830: INFO: (13) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 4.212179ms) Apr 25 00:42:02.830: INFO: (13) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.297126ms) Apr 25 00:42:02.830: INFO: (13) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 4.24325ms) Apr 25 00:42:02.830: INFO: (13) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.249981ms) Apr 25 00:42:02.830: INFO: (13) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 4.472859ms) Apr 25 00:42:02.830: INFO: (13) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 4.548837ms) Apr 25 00:42:02.831: INFO: (13) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.672873ms) Apr 25 00:42:02.831: INFO: (13) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 4.689156ms) Apr 25 00:42:02.831: INFO: (13) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 4.798769ms) Apr 25 00:42:02.831: INFO: (13) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.01588ms) Apr 25 00:42:02.831: INFO: (13) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.365602ms) Apr 25 00:42:02.841: INFO: (14) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 9.873356ms) Apr 25 00:42:02.841: INFO: (14) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 9.874313ms) Apr 25 00:42:02.842: INFO: (14) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 10.572045ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 11.218492ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 11.326098ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 11.506041ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 11.614041ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 11.605137ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 11.57359ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 11.561489ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 11.688684ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 11.680805ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 11.622453ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 11.748632ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 11.686262ms) Apr 25 00:42:02.843: INFO: (14) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test<... (200; 3.068915ms) Apr 25 00:42:02.846: INFO: (15) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 3.170704ms) Apr 25 00:42:02.847: INFO: (15) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 3.431231ms) Apr 25 00:42:02.847: INFO: (15) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 3.449931ms) Apr 25 00:42:02.849: INFO: (15) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.237465ms) Apr 25 00:42:02.849: INFO: (15) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 5.553556ms) Apr 25 00:42:02.849: INFO: (15) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: ... (200; 5.728213ms) Apr 25 00:42:02.849: INFO: (15) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.735048ms) Apr 25 00:42:02.850: INFO: (15) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 6.228905ms) Apr 25 00:42:02.850: INFO: (15) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 6.415179ms) Apr 25 00:42:02.850: INFO: (15) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 6.484452ms) Apr 25 00:42:02.850: INFO: (15) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 6.509372ms) Apr 25 00:42:02.850: INFO: (15) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 6.649536ms) Apr 25 00:42:02.850: INFO: (15) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 6.938682ms) Apr 25 00:42:02.852: INFO: (16) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 2.019815ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.154824ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.415285ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.528817ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 5.04895ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 5.12426ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.162786ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.166325ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.086698ms) Apr 25 00:42:02.855: INFO: (16) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 5.070412ms) Apr 25 00:42:02.856: INFO: (16) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.159915ms) Apr 25 00:42:02.856: INFO: (16) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 5.162161ms) Apr 25 00:42:02.856: INFO: (16) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 5.295456ms) Apr 25 00:42:02.856: INFO: (16) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.300287ms) Apr 25 00:42:02.856: INFO: (16) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 5.21292ms) Apr 25 00:42:02.856: INFO: (16) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test (200; 2.762859ms) Apr 25 00:42:02.859: INFO: (17) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 2.899911ms) Apr 25 00:42:02.859: INFO: (17) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 2.922771ms) Apr 25 00:42:02.859: INFO: (17) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test<... (200; 3.080848ms) Apr 25 00:42:02.860: INFO: (17) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 4.197778ms) Apr 25 00:42:02.860: INFO: (17) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 4.437246ms) Apr 25 00:42:02.860: INFO: (17) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 4.385402ms) Apr 25 00:42:02.860: INFO: (17) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 4.472563ms) Apr 25 00:42:02.860: INFO: (17) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 4.505339ms) Apr 25 00:42:02.860: INFO: (17) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.502761ms) Apr 25 00:42:02.861: INFO: (17) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 4.91079ms) Apr 25 00:42:02.861: INFO: (17) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.025928ms) Apr 25 00:42:02.861: INFO: (17) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.045273ms) Apr 25 00:42:02.862: INFO: (17) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 5.954513ms) Apr 25 00:42:02.862: INFO: (17) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.986071ms) Apr 25 00:42:02.865: INFO: (18) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:1080/proxy/: test<... (200; 3.109939ms) Apr 25 00:42:02.865: INFO: (18) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 3.388293ms) Apr 25 00:42:02.865: INFO: (18) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 3.537304ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 3.879443ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 3.880087ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:162/proxy/: bar (200; 3.869604ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 3.875454ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 4.074054ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt/proxy/: test (200; 4.201297ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 4.283663ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 4.246642ms) Apr 25 00:42:02.866: INFO: (18) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test<... (200; 4.934501ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/pods/proxy-service-wvcgb-rs6tt:160/proxy/: foo (200; 5.013841ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:443/proxy/: test (200; 5.087784ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname1/proxy/: tls baz (200; 5.05515ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:462/proxy/: tls qux (200; 5.06463ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/services/https:proxy-service-wvcgb:tlsportname2/proxy/: tls qux (200; 5.046249ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname2/proxy/: bar (200; 5.091657ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/pods/https:proxy-service-wvcgb-rs6tt:460/proxy/: tls baz (200; 5.149644ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname2/proxy/: bar (200; 5.166206ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/services/proxy-service-wvcgb:portname1/proxy/: foo (200; 5.055399ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/pods/http:proxy-service-wvcgb-rs6tt:1080/proxy/: ... (200; 5.115201ms) Apr 25 00:42:02.872: INFO: (19) /api/v1/namespaces/proxy-2545/services/http:proxy-service-wvcgb:portname1/proxy/: foo (200; 5.174574ms) STEP: deleting ReplicationController proxy-service-wvcgb in namespace proxy-2545, will wait for the garbage collector to delete the pods Apr 25 00:42:02.931: INFO: Deleting ReplicationController proxy-service-wvcgb took: 6.594446ms Apr 25 00:42:03.231: INFO: Terminating ReplicationController proxy-service-wvcgb pods took: 300.259973ms [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:05.631: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-2545" for this suite. • [SLOW TEST:15.383 seconds] [sig-network] Proxy /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:59 should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance]","total":275,"completed":204,"skipped":3701,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:05.641: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [BeforeEach] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1418 [It] should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: running the image docker.io/library/httpd:2.4.38-alpine Apr 25 00:42:05.764: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config run e2e-test-httpd-pod --restart=Never --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-7881' Apr 25 00:42:08.269: INFO: stderr: "" Apr 25 00:42:08.269: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: verifying the pod e2e-test-httpd-pod was created [AfterEach] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1423 Apr 25 00:42:08.290: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete pods e2e-test-httpd-pod --namespace=kubectl-7881' Apr 25 00:42:12.744: INFO: stderr: "" Apr 25 00:42:12.744: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:12.744: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7881" for this suite. • [SLOW TEST:7.146 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1414 should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance]","total":275,"completed":205,"skipped":3717,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:12.788: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:42:12.833: INFO: Waiting up to 5m0s for pod "downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4" in namespace "downward-api-2814" to be "Succeeded or Failed" Apr 25 00:42:12.841: INFO: Pod "downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4": Phase="Pending", Reason="", readiness=false. Elapsed: 8.163094ms Apr 25 00:42:14.844: INFO: Pod "downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011564282s Apr 25 00:42:16.849: INFO: Pod "downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016455354s STEP: Saw pod success Apr 25 00:42:16.849: INFO: Pod "downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4" satisfied condition "Succeeded or Failed" Apr 25 00:42:16.852: INFO: Trying to get logs from node latest-worker pod downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4 container client-container: STEP: delete the pod Apr 25 00:42:16.869: INFO: Waiting for pod downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4 to disappear Apr 25 00:42:16.889: INFO: Pod downwardapi-volume-69b10555-0e87-4cc2-a1a2-e20c2cd8dbe4 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:16.889: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-2814" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":275,"completed":206,"skipped":3732,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:16.898: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir volume type on tmpfs Apr 25 00:42:16.946: INFO: Waiting up to 5m0s for pod "pod-8a954aea-37a9-45b3-9c95-dc4804283151" in namespace "emptydir-591" to be "Succeeded or Failed" Apr 25 00:42:16.958: INFO: Pod "pod-8a954aea-37a9-45b3-9c95-dc4804283151": Phase="Pending", Reason="", readiness=false. Elapsed: 11.331926ms Apr 25 00:42:18.961: INFO: Pod "pod-8a954aea-37a9-45b3-9c95-dc4804283151": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015058215s Apr 25 00:42:20.965: INFO: Pod "pod-8a954aea-37a9-45b3-9c95-dc4804283151": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018938457s STEP: Saw pod success Apr 25 00:42:20.965: INFO: Pod "pod-8a954aea-37a9-45b3-9c95-dc4804283151" satisfied condition "Succeeded or Failed" Apr 25 00:42:20.968: INFO: Trying to get logs from node latest-worker2 pod pod-8a954aea-37a9-45b3-9c95-dc4804283151 container test-container: STEP: delete the pod Apr 25 00:42:20.995: INFO: Waiting for pod pod-8a954aea-37a9-45b3-9c95-dc4804283151 to disappear Apr 25 00:42:21.000: INFO: Pod pod-8a954aea-37a9-45b3-9c95-dc4804283151 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:21.000: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-591" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":207,"skipped":3752,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:21.007: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:178 [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:42:21.085: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:25.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8977" for this suite. •{"msg":"PASSED [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":275,"completed":208,"skipped":3773,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:25.147: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:178 [It] should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:42:25.189: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:31.323: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-3126" for this suite. • [SLOW TEST:6.186 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":275,"completed":209,"skipped":3806,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:31.333: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a new configmap STEP: modifying the configmap once STEP: modifying the configmap a second time STEP: deleting the configmap STEP: creating a watch on configmaps from the resource version returned by the first update STEP: Expecting to observe notifications for all changes to the configmap after the first update Apr 25 00:42:31.407: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-6430 /api/v1/namespaces/watch-6430/configmaps/e2e-watch-test-resource-version 679727a6-091e-420e-ac2e-72df0859538c 10799633 0 2020-04-25 00:42:31 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Apr 25 00:42:31.407: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-6430 /api/v1/namespaces/watch-6430/configmaps/e2e-watch-test-resource-version 679727a6-091e-420e-ac2e-72df0859538c 10799634 0 2020-04-25 00:42:31 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:31.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-6430" for this suite. •{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","total":275,"completed":210,"skipped":3825,"failed":0} SSSS ------------------------------ [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:31.413: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating Agnhost RC Apr 25 00:42:31.464: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-5500' Apr 25 00:42:31.749: INFO: stderr: "" Apr 25 00:42:31.750: INFO: stdout: "replicationcontroller/agnhost-master created\n" STEP: Waiting for Agnhost master to start. Apr 25 00:42:32.754: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:42:32.754: INFO: Found 0 / 1 Apr 25 00:42:33.754: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:42:33.754: INFO: Found 0 / 1 Apr 25 00:42:34.754: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:42:34.754: INFO: Found 1 / 1 Apr 25 00:42:34.754: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 STEP: patching all pods Apr 25 00:42:34.758: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:42:34.758: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Apr 25 00:42:34.758: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config patch pod agnhost-master-xs64c --namespace=kubectl-5500 -p {"metadata":{"annotations":{"x":"y"}}}' Apr 25 00:42:34.865: INFO: stderr: "" Apr 25 00:42:34.865: INFO: stdout: "pod/agnhost-master-xs64c patched\n" STEP: checking annotations Apr 25 00:42:34.868: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:42:34.868: INFO: ForEach: Found 1 pods from the filter. Now looping through them. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:34.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5500" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance]","total":275,"completed":211,"skipped":3829,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:34.874: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:38.955: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-1335" for this suite. •{"msg":"PASSED [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","total":275,"completed":212,"skipped":3846,"failed":0} ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:38.962: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0666 on tmpfs Apr 25 00:42:39.044: INFO: Waiting up to 5m0s for pod "pod-231b0faa-c55d-4841-8de5-adadc430a046" in namespace "emptydir-7899" to be "Succeeded or Failed" Apr 25 00:42:39.048: INFO: Pod "pod-231b0faa-c55d-4841-8de5-adadc430a046": Phase="Pending", Reason="", readiness=false. Elapsed: 3.514821ms Apr 25 00:42:41.073: INFO: Pod "pod-231b0faa-c55d-4841-8de5-adadc430a046": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028780955s Apr 25 00:42:43.076: INFO: Pod "pod-231b0faa-c55d-4841-8de5-adadc430a046": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.032421084s STEP: Saw pod success Apr 25 00:42:43.076: INFO: Pod "pod-231b0faa-c55d-4841-8de5-adadc430a046" satisfied condition "Succeeded or Failed" Apr 25 00:42:43.079: INFO: Trying to get logs from node latest-worker pod pod-231b0faa-c55d-4841-8de5-adadc430a046 container test-container: STEP: delete the pod Apr 25 00:42:43.175: INFO: Waiting for pod pod-231b0faa-c55d-4841-8de5-adadc430a046 to disappear Apr 25 00:42:43.180: INFO: Pod pod-231b0faa-c55d-4841-8de5-adadc430a046 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:43.180: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7899" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":213,"skipped":3846,"failed":0} S ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:43.193: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:91 Apr 25 00:42:43.302: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 25 00:42:43.312: INFO: Waiting for terminating namespaces to be deleted... Apr 25 00:42:43.315: INFO: Logging pods the kubelet thinks is on node latest-worker before test Apr 25 00:42:43.321: INFO: kube-proxy-s9v6p from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.321: INFO: Container kube-proxy ready: true, restart count 0 Apr 25 00:42:43.321: INFO: client-containers-f981b48e-e0f2-4b3d-8aaa-c45289f976a9 from containers-1335 started at 2020-04-25 00:42:34 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.321: INFO: Container test-container ready: true, restart count 0 Apr 25 00:42:43.321: INFO: agnhost-master-xs64c from kubectl-5500 started at 2020-04-25 00:42:31 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.321: INFO: Container agnhost-master ready: false, restart count 0 Apr 25 00:42:43.321: INFO: kindnet-vnjgh from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.321: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:42:43.322: INFO: Logging pods the kubelet thinks is on node latest-worker2 before test Apr 25 00:42:43.327: INFO: pod-exec-websocket-eef4b07b-6e2a-4583-9e78-ede66d5a6d06 from pods-3126 started at 2020-04-25 00:42:25 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.327: INFO: Container main ready: true, restart count 0 Apr 25 00:42:43.327: INFO: pod-logs-websocket-61731a41-18cd-439c-8f6a-201c4c4cee6e from pods-8977 started at 2020-04-25 00:42:21 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.327: INFO: Container main ready: true, restart count 0 Apr 25 00:42:43.327: INFO: kindnet-zq6gp from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.327: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:42:43.327: INFO: kube-proxy-c5xlk from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:42:43.327: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Trying to schedule Pod with nonempty NodeSelector. STEP: Considering event: Type = [Warning], Name = [restricted-pod.1608e947d32ecbbb], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.] [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:42:44.387: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-5265" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:82 •{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance]","total":275,"completed":214,"skipped":3847,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:42:44.395: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod pod-subpath-test-downwardapi-zzxc STEP: Creating a pod to test atomic-volume-subpath Apr 25 00:42:44.544: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-zzxc" in namespace "subpath-4910" to be "Succeeded or Failed" Apr 25 00:42:44.554: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Pending", Reason="", readiness=false. Elapsed: 9.492208ms Apr 25 00:42:46.557: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013291378s Apr 25 00:42:48.560: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 4.015987714s Apr 25 00:42:50.564: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 6.020296708s Apr 25 00:42:52.568: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 8.024145997s Apr 25 00:42:54.572: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 10.028222297s Apr 25 00:42:56.578: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 12.034344101s Apr 25 00:42:58.581: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 14.037168791s Apr 25 00:43:00.585: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 16.041149235s Apr 25 00:43:02.588: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 18.044455125s Apr 25 00:43:04.593: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 20.048820256s Apr 25 00:43:06.596: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Running", Reason="", readiness=true. Elapsed: 22.052162238s Apr 25 00:43:08.599: INFO: Pod "pod-subpath-test-downwardapi-zzxc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.055314319s STEP: Saw pod success Apr 25 00:43:08.599: INFO: Pod "pod-subpath-test-downwardapi-zzxc" satisfied condition "Succeeded or Failed" Apr 25 00:43:08.602: INFO: Trying to get logs from node latest-worker pod pod-subpath-test-downwardapi-zzxc container test-container-subpath-downwardapi-zzxc: STEP: delete the pod Apr 25 00:43:08.632: INFO: Waiting for pod pod-subpath-test-downwardapi-zzxc to disappear Apr 25 00:43:08.662: INFO: Pod pod-subpath-test-downwardapi-zzxc no longer exists STEP: Deleting pod pod-subpath-test-downwardapi-zzxc Apr 25 00:43:08.662: INFO: Deleting pod "pod-subpath-test-downwardapi-zzxc" in namespace "subpath-4910" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:08.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-4910" for this suite. • [SLOW TEST:24.276 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]","total":275,"completed":215,"skipped":3868,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:08.673: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:43:08.734: INFO: (0) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 6.911806ms) Apr 25 00:43:08.737: INFO: (1) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.803164ms) Apr 25 00:43:08.739: INFO: (2) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.665051ms) Apr 25 00:43:08.742: INFO: (3) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.652632ms) Apr 25 00:43:08.744: INFO: (4) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.498537ms) Apr 25 00:43:08.747: INFO: (5) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.626651ms) Apr 25 00:43:08.750: INFO: (6) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.900331ms) Apr 25 00:43:08.752: INFO: (7) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.269066ms) Apr 25 00:43:08.755: INFO: (8) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.770501ms) Apr 25 00:43:08.758: INFO: (9) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.451206ms) Apr 25 00:43:08.760: INFO: (10) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.108797ms) Apr 25 00:43:08.762: INFO: (11) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.426227ms) Apr 25 00:43:08.783: INFO: (12) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 20.637875ms) Apr 25 00:43:08.786: INFO: (13) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.019647ms) Apr 25 00:43:08.789: INFO: (14) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.184813ms) Apr 25 00:43:08.792: INFO: (15) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.502401ms) Apr 25 00:43:08.794: INFO: (16) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.732096ms) Apr 25 00:43:08.798: INFO: (17) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.067191ms) Apr 25 00:43:08.801: INFO: (18) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.241181ms) Apr 25 00:43:08.806: INFO: (19) /api/v1/nodes/latest-worker:10250/proxy/logs/:
containers/
pods/
(200; 5.35905ms) [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:08.806: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-9158" for this suite. •{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource [Conformance]","total":275,"completed":216,"skipped":3931,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:08.814: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-16bfe1d9-de09-4636-937f-f15852592fdd STEP: Creating a pod to test consume secrets Apr 25 00:43:08.904: INFO: Waiting up to 5m0s for pod "pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525" in namespace "secrets-2249" to be "Succeeded or Failed" Apr 25 00:43:08.947: INFO: Pod "pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525": Phase="Pending", Reason="", readiness=false. Elapsed: 42.907377ms Apr 25 00:43:10.951: INFO: Pod "pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04721964s Apr 25 00:43:12.956: INFO: Pod "pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.051566791s STEP: Saw pod success Apr 25 00:43:12.956: INFO: Pod "pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525" satisfied condition "Succeeded or Failed" Apr 25 00:43:12.959: INFO: Trying to get logs from node latest-worker pod pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525 container secret-volume-test: STEP: delete the pod Apr 25 00:43:13.016: INFO: Waiting for pod pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525 to disappear Apr 25 00:43:13.019: INFO: Pod pod-secrets-b00703b9-8396-47ba-8574-cdee698d1525 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:13.019: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-2249" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":217,"skipped":3989,"failed":0} SS ------------------------------ [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:13.026: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for all rs to be garbage collected STEP: expected 0 rs, got 1 rs STEP: expected 0 pods, got 2 pods STEP: Gathering metrics W0425 00:43:14.257538 8 metrics_grabber.go:84] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 25 00:43:14.257: INFO: For apiserver_request_total: For apiserver_request_latency_seconds: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:14.257: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-6508" for this suite. •{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":275,"completed":218,"skipped":3991,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:14.284: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name projected-configmap-test-volume-6af562f1-6249-4ebb-a73d-38feac1a80e7 STEP: Creating a pod to test consume configMaps Apr 25 00:43:14.413: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85" in namespace "projected-644" to be "Succeeded or Failed" Apr 25 00:43:14.417: INFO: Pod "pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85": Phase="Pending", Reason="", readiness=false. Elapsed: 4.216685ms Apr 25 00:43:16.441: INFO: Pod "pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028264132s Apr 25 00:43:18.445: INFO: Pod "pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.032177933s STEP: Saw pod success Apr 25 00:43:18.445: INFO: Pod "pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85" satisfied condition "Succeeded or Failed" Apr 25 00:43:18.448: INFO: Trying to get logs from node latest-worker pod pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85 container projected-configmap-volume-test: STEP: delete the pod Apr 25 00:43:18.501: INFO: Waiting for pod pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85 to disappear Apr 25 00:43:18.513: INFO: Pod pod-projected-configmaps-aa14ce1b-9ea8-4fdd-ab9f-be71c9e63f85 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:18.513: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-644" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":219,"skipped":4006,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:18.520: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:43:19.336: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:43:21.346: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372199, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372199, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372199, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372199, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:43:24.362: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] listing validating webhooks should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Listing all of the created validation webhooks STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Deleting the collection of validation webhooks STEP: Creating a configMap that does not comply to the validation webhook rules [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:25.245: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-4754" for this suite. STEP: Destroying namespace "webhook-4754-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:6.821 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 listing validating webhooks should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","total":275,"completed":220,"skipped":4014,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:25.342: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating projection with secret that has name secret-emptykey-test-87d1c4da-149c-46fa-bd49-623024389ced [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:25.469: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-59" for this suite. •{"msg":"PASSED [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance]","total":275,"completed":221,"skipped":4064,"failed":0} SSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:25.476: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: starting the proxy server Apr 25 00:43:25.888: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config proxy -p 0 --disable-filter' STEP: curling proxy /api/ output [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:25.983: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4022" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]","total":275,"completed":222,"skipped":4075,"failed":0} SSSSSSSSSSS ------------------------------ [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:25.992: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:178 [It] should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod STEP: setting up watch STEP: submitting the pod to kubernetes Apr 25 00:43:26.174: INFO: observed the pod list STEP: verifying the pod is in kubernetes STEP: verifying pod creation was observed STEP: deleting the pod gracefully STEP: verifying the kubelet observed the termination notice STEP: verifying pod deletion was observed [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:43.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-2214" for this suite. • [SLOW TEST:17.024 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance]","total":275,"completed":223,"skipped":4086,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:43.016: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating projection with secret that has name projected-secret-test-db559d3e-d4d3-4ffe-97ed-4c5d456f5657 STEP: Creating a pod to test consume secrets Apr 25 00:43:43.082: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a" in namespace "projected-3033" to be "Succeeded or Failed" Apr 25 00:43:43.086: INFO: Pod "pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a": Phase="Pending", Reason="", readiness=false. Elapsed: 3.887159ms Apr 25 00:43:45.090: INFO: Pod "pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008081212s Apr 25 00:43:47.095: INFO: Pod "pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012629359s STEP: Saw pod success Apr 25 00:43:47.095: INFO: Pod "pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a" satisfied condition "Succeeded or Failed" Apr 25 00:43:47.098: INFO: Trying to get logs from node latest-worker2 pod pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a container projected-secret-volume-test: STEP: delete the pod Apr 25 00:43:47.136: INFO: Waiting for pod pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a to disappear Apr 25 00:43:47.138: INFO: Pod pod-projected-secrets-d98e1e59-29f9-4cb8-9295-246b5432ac0a no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:47.139: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3033" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":224,"skipped":4103,"failed":0} SSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:47.146: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:43:47.736: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:43:49.747: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:43:51.759: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372227, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:43:54.777: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] listing mutating webhooks should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Listing all of the created validation webhooks STEP: Creating a configMap that should be mutated STEP: Deleting the collection of validation webhooks STEP: Creating a configMap that should not be mutated [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:43:55.215: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-8349" for this suite. STEP: Destroying namespace "webhook-8349-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:8.121 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 listing mutating webhooks should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","total":275,"completed":225,"skipped":4116,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to create a functioning NodePort service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:43:55.268: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:698 [It] should be able to create a functioning NodePort service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating service nodeport-test with type=NodePort in namespace services-2168 STEP: creating replication controller nodeport-test in namespace services-2168 I0425 00:43:55.430099 8 runners.go:190] Created replication controller with name: nodeport-test, namespace: services-2168, replica count: 2 I0425 00:43:58.480622 8 runners.go:190] nodeport-test Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0425 00:44:01.480884 8 runners.go:190] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Apr 25 00:44:01.480: INFO: Creating new exec pod Apr 25 00:44:08.499: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-2168 execpodfzsn4 -- /bin/sh -x -c nc -zv -t -w 2 nodeport-test 80' Apr 25 00:44:08.692: INFO: stderr: "I0425 00:44:08.629807 2562 log.go:172] (0xc00020dce0) (0xc000549360) Create stream\nI0425 00:44:08.629862 2562 log.go:172] (0xc00020dce0) (0xc000549360) Stream added, broadcasting: 1\nI0425 00:44:08.632049 2562 log.go:172] (0xc00020dce0) Reply frame received for 1\nI0425 00:44:08.632091 2562 log.go:172] (0xc00020dce0) (0xc000549400) Create stream\nI0425 00:44:08.632099 2562 log.go:172] (0xc00020dce0) (0xc000549400) Stream added, broadcasting: 3\nI0425 00:44:08.633017 2562 log.go:172] (0xc00020dce0) Reply frame received for 3\nI0425 00:44:08.633047 2562 log.go:172] (0xc00020dce0) (0xc00071c000) Create stream\nI0425 00:44:08.633059 2562 log.go:172] (0xc00020dce0) (0xc00071c000) Stream added, broadcasting: 5\nI0425 00:44:08.633929 2562 log.go:172] (0xc00020dce0) Reply frame received for 5\nI0425 00:44:08.683832 2562 log.go:172] (0xc00020dce0) Data frame received for 5\nI0425 00:44:08.683934 2562 log.go:172] (0xc00071c000) (5) Data frame handling\nI0425 00:44:08.683977 2562 log.go:172] (0xc00071c000) (5) Data frame sent\n+ nc -zv -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\nI0425 00:44:08.684282 2562 log.go:172] (0xc00020dce0) Data frame received for 3\nI0425 00:44:08.684365 2562 log.go:172] (0xc000549400) (3) Data frame handling\nI0425 00:44:08.684444 2562 log.go:172] (0xc00020dce0) Data frame received for 5\nI0425 00:44:08.684471 2562 log.go:172] (0xc00071c000) (5) Data frame handling\nI0425 00:44:08.686938 2562 log.go:172] (0xc00020dce0) Data frame received for 1\nI0425 00:44:08.686982 2562 log.go:172] (0xc000549360) (1) Data frame handling\nI0425 00:44:08.687013 2562 log.go:172] (0xc000549360) (1) Data frame sent\nI0425 00:44:08.687051 2562 log.go:172] (0xc00020dce0) (0xc000549360) Stream removed, broadcasting: 1\nI0425 00:44:08.687093 2562 log.go:172] (0xc00020dce0) Go away received\nI0425 00:44:08.687566 2562 log.go:172] (0xc00020dce0) (0xc000549360) Stream removed, broadcasting: 1\nI0425 00:44:08.687591 2562 log.go:172] (0xc00020dce0) (0xc000549400) Stream removed, broadcasting: 3\nI0425 00:44:08.687604 2562 log.go:172] (0xc00020dce0) (0xc00071c000) Stream removed, broadcasting: 5\n" Apr 25 00:44:08.692: INFO: stdout: "" Apr 25 00:44:08.693: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-2168 execpodfzsn4 -- /bin/sh -x -c nc -zv -t -w 2 10.96.207.76 80' Apr 25 00:44:08.879: INFO: stderr: "I0425 00:44:08.801397 2583 log.go:172] (0xc000a76630) (0xc000a00780) Create stream\nI0425 00:44:08.801453 2583 log.go:172] (0xc000a76630) (0xc000a00780) Stream added, broadcasting: 1\nI0425 00:44:08.805060 2583 log.go:172] (0xc000a76630) Reply frame received for 1\nI0425 00:44:08.805239 2583 log.go:172] (0xc000a76630) (0xc000a506e0) Create stream\nI0425 00:44:08.805254 2583 log.go:172] (0xc000a76630) (0xc000a506e0) Stream added, broadcasting: 3\nI0425 00:44:08.806262 2583 log.go:172] (0xc000a76630) Reply frame received for 3\nI0425 00:44:08.806316 2583 log.go:172] (0xc000a76630) (0xc000a00000) Create stream\nI0425 00:44:08.806333 2583 log.go:172] (0xc000a76630) (0xc000a00000) Stream added, broadcasting: 5\nI0425 00:44:08.807350 2583 log.go:172] (0xc000a76630) Reply frame received for 5\nI0425 00:44:08.872338 2583 log.go:172] (0xc000a76630) Data frame received for 5\nI0425 00:44:08.872402 2583 log.go:172] (0xc000a00000) (5) Data frame handling\nI0425 00:44:08.872421 2583 log.go:172] (0xc000a00000) (5) Data frame sent\nI0425 00:44:08.872443 2583 log.go:172] (0xc000a76630) Data frame received for 5\nI0425 00:44:08.872462 2583 log.go:172] (0xc000a00000) (5) Data frame handling\n+ nc -zv -t -w 2 10.96.207.76 80\nConnection to 10.96.207.76 80 port [tcp/http] succeeded!\nI0425 00:44:08.872503 2583 log.go:172] (0xc000a76630) Data frame received for 3\nI0425 00:44:08.872531 2583 log.go:172] (0xc000a506e0) (3) Data frame handling\nI0425 00:44:08.874276 2583 log.go:172] (0xc000a76630) Data frame received for 1\nI0425 00:44:08.874305 2583 log.go:172] (0xc000a00780) (1) Data frame handling\nI0425 00:44:08.874318 2583 log.go:172] (0xc000a00780) (1) Data frame sent\nI0425 00:44:08.874335 2583 log.go:172] (0xc000a76630) (0xc000a00780) Stream removed, broadcasting: 1\nI0425 00:44:08.874602 2583 log.go:172] (0xc000a76630) Go away received\nI0425 00:44:08.874752 2583 log.go:172] (0xc000a76630) (0xc000a00780) Stream removed, broadcasting: 1\nI0425 00:44:08.874771 2583 log.go:172] (0xc000a76630) (0xc000a506e0) Stream removed, broadcasting: 3\nI0425 00:44:08.874782 2583 log.go:172] (0xc000a76630) (0xc000a00000) Stream removed, broadcasting: 5\n" Apr 25 00:44:08.879: INFO: stdout: "" Apr 25 00:44:08.879: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-2168 execpodfzsn4 -- /bin/sh -x -c nc -zv -t -w 2 172.17.0.13 32061' Apr 25 00:44:09.084: INFO: stderr: "I0425 00:44:09.008619 2603 log.go:172] (0xc000b15080) (0xc0009c8780) Create stream\nI0425 00:44:09.008679 2603 log.go:172] (0xc000b15080) (0xc0009c8780) Stream added, broadcasting: 1\nI0425 00:44:09.013371 2603 log.go:172] (0xc000b15080) Reply frame received for 1\nI0425 00:44:09.013440 2603 log.go:172] (0xc000b15080) (0xc0006a3720) Create stream\nI0425 00:44:09.013464 2603 log.go:172] (0xc000b15080) (0xc0006a3720) Stream added, broadcasting: 3\nI0425 00:44:09.014504 2603 log.go:172] (0xc000b15080) Reply frame received for 3\nI0425 00:44:09.014550 2603 log.go:172] (0xc000b15080) (0xc000524b40) Create stream\nI0425 00:44:09.014565 2603 log.go:172] (0xc000b15080) (0xc000524b40) Stream added, broadcasting: 5\nI0425 00:44:09.015614 2603 log.go:172] (0xc000b15080) Reply frame received for 5\nI0425 00:44:09.077797 2603 log.go:172] (0xc000b15080) Data frame received for 3\nI0425 00:44:09.077829 2603 log.go:172] (0xc0006a3720) (3) Data frame handling\nI0425 00:44:09.077895 2603 log.go:172] (0xc000b15080) Data frame received for 5\nI0425 00:44:09.077945 2603 log.go:172] (0xc000524b40) (5) Data frame handling\nI0425 00:44:09.077971 2603 log.go:172] (0xc000524b40) (5) Data frame sent\n+ nc -zv -t -w 2 172.17.0.13 32061\nConnection to 172.17.0.13 32061 port [tcp/32061] succeeded!\nI0425 00:44:09.078146 2603 log.go:172] (0xc000b15080) Data frame received for 5\nI0425 00:44:09.078179 2603 log.go:172] (0xc000524b40) (5) Data frame handling\nI0425 00:44:09.079502 2603 log.go:172] (0xc000b15080) Data frame received for 1\nI0425 00:44:09.079524 2603 log.go:172] (0xc0009c8780) (1) Data frame handling\nI0425 00:44:09.079536 2603 log.go:172] (0xc0009c8780) (1) Data frame sent\nI0425 00:44:09.079553 2603 log.go:172] (0xc000b15080) (0xc0009c8780) Stream removed, broadcasting: 1\nI0425 00:44:09.079570 2603 log.go:172] (0xc000b15080) Go away received\nI0425 00:44:09.080041 2603 log.go:172] (0xc000b15080) (0xc0009c8780) Stream removed, broadcasting: 1\nI0425 00:44:09.080061 2603 log.go:172] (0xc000b15080) (0xc0006a3720) Stream removed, broadcasting: 3\nI0425 00:44:09.080071 2603 log.go:172] (0xc000b15080) (0xc000524b40) Stream removed, broadcasting: 5\n" Apr 25 00:44:09.084: INFO: stdout: "" Apr 25 00:44:09.085: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config exec --namespace=services-2168 execpodfzsn4 -- /bin/sh -x -c nc -zv -t -w 2 172.17.0.12 32061' Apr 25 00:44:09.278: INFO: stderr: "I0425 00:44:09.213004 2624 log.go:172] (0xc0009f2000) (0xc0009ec000) Create stream\nI0425 00:44:09.213048 2624 log.go:172] (0xc0009f2000) (0xc0009ec000) Stream added, broadcasting: 1\nI0425 00:44:09.214859 2624 log.go:172] (0xc0009f2000) Reply frame received for 1\nI0425 00:44:09.214884 2624 log.go:172] (0xc0009f2000) (0xc000518b40) Create stream\nI0425 00:44:09.214891 2624 log.go:172] (0xc0009f2000) (0xc000518b40) Stream added, broadcasting: 3\nI0425 00:44:09.215605 2624 log.go:172] (0xc0009f2000) Reply frame received for 3\nI0425 00:44:09.215629 2624 log.go:172] (0xc0009f2000) (0xc0009ec0a0) Create stream\nI0425 00:44:09.215638 2624 log.go:172] (0xc0009f2000) (0xc0009ec0a0) Stream added, broadcasting: 5\nI0425 00:44:09.216221 2624 log.go:172] (0xc0009f2000) Reply frame received for 5\nI0425 00:44:09.270569 2624 log.go:172] (0xc0009f2000) Data frame received for 3\nI0425 00:44:09.270614 2624 log.go:172] (0xc000518b40) (3) Data frame handling\nI0425 00:44:09.270633 2624 log.go:172] (0xc0009f2000) Data frame received for 5\nI0425 00:44:09.270639 2624 log.go:172] (0xc0009ec0a0) (5) Data frame handling\nI0425 00:44:09.270646 2624 log.go:172] (0xc0009ec0a0) (5) Data frame sent\nI0425 00:44:09.270652 2624 log.go:172] (0xc0009f2000) Data frame received for 5\nI0425 00:44:09.270657 2624 log.go:172] (0xc0009ec0a0) (5) Data frame handling\n+ nc -zv -t -w 2 172.17.0.12 32061\nConnection to 172.17.0.12 32061 port [tcp/32061] succeeded!\nI0425 00:44:09.272315 2624 log.go:172] (0xc0009f2000) Data frame received for 1\nI0425 00:44:09.272354 2624 log.go:172] (0xc0009ec000) (1) Data frame handling\nI0425 00:44:09.272378 2624 log.go:172] (0xc0009ec000) (1) Data frame sent\nI0425 00:44:09.272403 2624 log.go:172] (0xc0009f2000) (0xc0009ec000) Stream removed, broadcasting: 1\nI0425 00:44:09.272429 2624 log.go:172] (0xc0009f2000) Go away received\nI0425 00:44:09.272746 2624 log.go:172] (0xc0009f2000) (0xc0009ec000) Stream removed, broadcasting: 1\nI0425 00:44:09.272771 2624 log.go:172] (0xc0009f2000) (0xc000518b40) Stream removed, broadcasting: 3\nI0425 00:44:09.272782 2624 log.go:172] (0xc0009f2000) (0xc0009ec0a0) Stream removed, broadcasting: 5\n" Apr 25 00:44:09.278: INFO: stdout: "" [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:44:09.278: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-2168" for this suite. [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:702 • [SLOW TEST:14.018 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to create a functioning NodePort service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","total":275,"completed":226,"skipped":4169,"failed":0} SSSS ------------------------------ [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:44:09.286: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [BeforeEach] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:271 [It] should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a replication controller Apr 25 00:44:09.360: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6316' Apr 25 00:44:09.618: INFO: stderr: "" Apr 25 00:44:09.618: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 25 00:44:09.618: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6316' Apr 25 00:44:09.754: INFO: stderr: "" Apr 25 00:44:09.754: INFO: stdout: "update-demo-nautilus-7sv6t update-demo-nautilus-n4ft9 " Apr 25 00:44:09.754: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7sv6t -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6316' Apr 25 00:44:09.843: INFO: stderr: "" Apr 25 00:44:09.843: INFO: stdout: "" Apr 25 00:44:09.843: INFO: update-demo-nautilus-7sv6t is created but not running Apr 25 00:44:14.843: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6316' Apr 25 00:44:14.974: INFO: stderr: "" Apr 25 00:44:14.974: INFO: stdout: "update-demo-nautilus-7sv6t update-demo-nautilus-n4ft9 " Apr 25 00:44:14.975: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7sv6t -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6316' Apr 25 00:44:15.063: INFO: stderr: "" Apr 25 00:44:15.063: INFO: stdout: "true" Apr 25 00:44:15.063: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7sv6t -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6316' Apr 25 00:44:15.149: INFO: stderr: "" Apr 25 00:44:15.150: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 25 00:44:15.150: INFO: validating pod update-demo-nautilus-7sv6t Apr 25 00:44:15.153: INFO: got data: { "image": "nautilus.jpg" } Apr 25 00:44:15.153: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 25 00:44:15.154: INFO: update-demo-nautilus-7sv6t is verified up and running Apr 25 00:44:15.154: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-n4ft9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6316' Apr 25 00:44:15.257: INFO: stderr: "" Apr 25 00:44:15.257: INFO: stdout: "true" Apr 25 00:44:15.257: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-n4ft9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6316' Apr 25 00:44:15.351: INFO: stderr: "" Apr 25 00:44:15.352: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 25 00:44:15.352: INFO: validating pod update-demo-nautilus-n4ft9 Apr 25 00:44:15.355: INFO: got data: { "image": "nautilus.jpg" } Apr 25 00:44:15.355: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 25 00:44:15.355: INFO: update-demo-nautilus-n4ft9 is verified up and running STEP: using delete to clean up resources Apr 25 00:44:15.355: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6316' Apr 25 00:44:15.458: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 25 00:44:15.458: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Apr 25 00:44:15.458: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-6316' Apr 25 00:44:15.542: INFO: stderr: "No resources found in kubectl-6316 namespace.\n" Apr 25 00:44:15.542: INFO: stdout: "" Apr 25 00:44:15.542: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-6316 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 25 00:44:15.632: INFO: stderr: "" Apr 25 00:44:15.632: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:44:15.632: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6316" for this suite. • [SLOW TEST:6.353 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:269 should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]","total":275,"completed":227,"skipped":4173,"failed":0} SSSSSSS ------------------------------ [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:44:15.639: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:44:15.724: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-dfdb977c-edb9-497e-bd7d-d994e9687db1" in namespace "security-context-test-2217" to be "Succeeded or Failed" Apr 25 00:44:15.766: INFO: Pod "alpine-nnp-false-dfdb977c-edb9-497e-bd7d-d994e9687db1": Phase="Pending", Reason="", readiness=false. Elapsed: 41.621394ms Apr 25 00:44:17.782: INFO: Pod "alpine-nnp-false-dfdb977c-edb9-497e-bd7d-d994e9687db1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.057515695s Apr 25 00:44:19.784: INFO: Pod "alpine-nnp-false-dfdb977c-edb9-497e-bd7d-d994e9687db1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.060075992s Apr 25 00:44:19.784: INFO: Pod "alpine-nnp-false-dfdb977c-edb9-497e-bd7d-d994e9687db1" satisfied condition "Succeeded or Failed" [AfterEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:44:19.789: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-2217" for this suite. •{"msg":"PASSED [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":228,"skipped":4180,"failed":0} SSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:44:19.796: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating configMap with name configmap-test-volume-map-3e5828f3-07ad-4da6-9db4-71f0313c183b STEP: Creating a pod to test consume configMaps Apr 25 00:44:19.867: INFO: Waiting up to 5m0s for pod "pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4" in namespace "configmap-6019" to be "Succeeded or Failed" Apr 25 00:44:19.884: INFO: Pod "pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4": Phase="Pending", Reason="", readiness=false. Elapsed: 16.438494ms Apr 25 00:44:21.888: INFO: Pod "pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020517914s Apr 25 00:44:23.892: INFO: Pod "pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024868066s STEP: Saw pod success Apr 25 00:44:23.892: INFO: Pod "pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4" satisfied condition "Succeeded or Failed" Apr 25 00:44:23.895: INFO: Trying to get logs from node latest-worker2 pod pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4 container configmap-volume-test: STEP: delete the pod Apr 25 00:44:23.927: INFO: Waiting for pod pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4 to disappear Apr 25 00:44:23.940: INFO: Pod pod-configmaps-56320212-c7df-439d-9d8b-68462aeaabf4 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:44:23.940: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-6019" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":275,"completed":229,"skipped":4190,"failed":0} SS ------------------------------ [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:44:23.949: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating the pod Apr 25 00:44:28.567: INFO: Successfully updated pod "annotationupdatedbef990e-40c4-4385-893c-1ee73bcb428f" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:44:30.585: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-3465" for this suite. • [SLOW TEST:6.647 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":275,"completed":230,"skipped":4192,"failed":0} SSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:44:30.596: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0644 on tmpfs Apr 25 00:44:30.664: INFO: Waiting up to 5m0s for pod "pod-868f50db-00e9-452b-b288-51d6deb4e6ce" in namespace "emptydir-1198" to be "Succeeded or Failed" Apr 25 00:44:30.667: INFO: Pod "pod-868f50db-00e9-452b-b288-51d6deb4e6ce": Phase="Pending", Reason="", readiness=false. Elapsed: 3.739937ms Apr 25 00:44:32.690: INFO: Pod "pod-868f50db-00e9-452b-b288-51d6deb4e6ce": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026149923s Apr 25 00:44:34.694: INFO: Pod "pod-868f50db-00e9-452b-b288-51d6deb4e6ce": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.03086329s STEP: Saw pod success Apr 25 00:44:34.695: INFO: Pod "pod-868f50db-00e9-452b-b288-51d6deb4e6ce" satisfied condition "Succeeded or Failed" Apr 25 00:44:34.698: INFO: Trying to get logs from node latest-worker pod pod-868f50db-00e9-452b-b288-51d6deb4e6ce container test-container: STEP: delete the pod Apr 25 00:44:34.767: INFO: Waiting for pod pod-868f50db-00e9-452b-b288-51d6deb4e6ce to disappear Apr 25 00:44:34.776: INFO: Pod pod-868f50db-00e9-452b-b288-51d6deb4e6ce no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:44:34.776: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1198" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":231,"skipped":4201,"failed":0} ------------------------------ [sig-apps] Deployment deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:44:34.783: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:74 [It] deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:44:34.854: INFO: Pod name rollover-pod: Found 0 pods out of 1 Apr 25 00:44:39.857: INFO: Pod name rollover-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Apr 25 00:44:39.857: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready Apr 25 00:44:41.860: INFO: Creating deployment "test-rollover-deployment" Apr 25 00:44:41.869: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations Apr 25 00:44:43.875: INFO: Check revision of new replica set for deployment "test-rollover-deployment" Apr 25 00:44:43.882: INFO: Ensure that both replica sets have 1 created replica Apr 25 00:44:43.888: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update Apr 25 00:44:43.894: INFO: Updating deployment test-rollover-deployment Apr 25 00:44:43.894: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller Apr 25 00:44:45.903: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 Apr 25 00:44:45.909: INFO: Make sure deployment "test-rollover-deployment" is complete Apr 25 00:44:45.915: INFO: all replica sets need to contain the pod-template-hash label Apr 25 00:44:45.915: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372284, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-78df7bc796\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:44:47.923: INFO: all replica sets need to contain the pod-template-hash label Apr 25 00:44:47.923: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372286, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-78df7bc796\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:44:49.923: INFO: all replica sets need to contain the pod-template-hash label Apr 25 00:44:49.923: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372286, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-78df7bc796\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:44:51.923: INFO: all replica sets need to contain the pod-template-hash label Apr 25 00:44:51.923: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372286, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-78df7bc796\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:44:53.923: INFO: all replica sets need to contain the pod-template-hash label Apr 25 00:44:53.923: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372286, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-78df7bc796\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:44:55.923: INFO: all replica sets need to contain the pod-template-hash label Apr 25 00:44:55.923: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372286, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372281, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-78df7bc796\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:44:57.923: INFO: Apr 25 00:44:57.923: INFO: Ensure that both old replica sets have no replicas [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:68 Apr 25 00:44:57.932: INFO: Deployment "test-rollover-deployment": &Deployment{ObjectMeta:{test-rollover-deployment deployment-2590 /apis/apps/v1/namespaces/deployment-2590/deployments/test-rollover-deployment 27b06950-05a5-4d46-bd8d-97a8f6f0b377 10800854 2 2020-04-25 00:44:41 +0000 UTC map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] [] []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod] map[] [] [] []} {[] [] [{agnhost us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0022cdb88 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-04-25 00:44:41 +0000 UTC,LastTransitionTime:2020-04-25 00:44:41 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-78df7bc796" has successfully progressed.,LastUpdateTime:2020-04-25 00:44:56 +0000 UTC,LastTransitionTime:2020-04-25 00:44:41 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Apr 25 00:44:57.935: INFO: New ReplicaSet "test-rollover-deployment-78df7bc796" of Deployment "test-rollover-deployment": &ReplicaSet{ObjectMeta:{test-rollover-deployment-78df7bc796 deployment-2590 /apis/apps/v1/namespaces/deployment-2590/replicasets/test-rollover-deployment-78df7bc796 7df4431b-bfc6-41c5-8fdc-02237f6a1198 10800843 2 2020-04-25 00:44:43 +0000 UTC map[name:rollover-pod pod-template-hash:78df7bc796] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment 27b06950-05a5-4d46-bd8d-97a8f6f0b377 0xc0027bc137 0xc0027bc138}] [] []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 78df7bc796,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:78df7bc796] map[] [] [] []} {[] [] [{agnhost us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0027bc1a8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:44:57.935: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": Apr 25 00:44:57.936: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller deployment-2590 /apis/apps/v1/namespaces/deployment-2590/replicasets/test-rollover-controller 0529e39d-110c-44e4-97ad-15ad4de5c97b 10800852 2 2020-04-25 00:44:34 +0000 UTC map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment 27b06950-05a5-4d46-bd8d-97a8f6f0b377 0xc0027bc067 0xc0027bc068}] [] []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod:httpd] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0027bc0c8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:44:57.936: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-f6c94f66c deployment-2590 /apis/apps/v1/namespaces/deployment-2590/replicasets/test-rollover-deployment-f6c94f66c 734ecf82-9eed-42d1-8737-ce2e7f5f1992 10800793 2 2020-04-25 00:44:41 +0000 UTC map[name:rollover-pod pod-template-hash:f6c94f66c] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment 27b06950-05a5-4d46-bd8d-97a8f6f0b377 0xc0027bc210 0xc0027bc211}] [] []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: f6c94f66c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:f6c94f66c] map[] [] [] []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0027bc338 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:44:57.939: INFO: Pod "test-rollover-deployment-78df7bc796-p9wc8" is available: &Pod{ObjectMeta:{test-rollover-deployment-78df7bc796-p9wc8 test-rollover-deployment-78df7bc796- deployment-2590 /api/v1/namespaces/deployment-2590/pods/test-rollover-deployment-78df7bc796-p9wc8 83fa4b7a-459c-4519-a806-5079a765d13b 10800811 0 2020-04-25 00:44:43 +0000 UTC map[name:rollover-pod pod-template-hash:78df7bc796] map[] [{apps/v1 ReplicaSet test-rollover-deployment-78df7bc796 7df4431b-bfc6-41c5-8fdc-02237f6a1198 0xc002ca4377 0xc002ca4378}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-gn6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-gn6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-gn6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:44:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:44:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:44:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:44:44 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.13,PodIP:10.244.2.246,StartTime:2020-04-25 00:44:44 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:44:46 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,ImageID:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost@sha256:1d7f0d77a6f07fd507f147a38d06a7c8269ebabd4f923bfe46d4fb8b396a520c,ContainerID:containerd://c2300cef1f953badecb235f7be3ae3addc7337f5425572fb1286a1f5f70a97d2,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.246,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:44:57.939: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-2590" for this suite. • [SLOW TEST:23.164 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","total":275,"completed":232,"skipped":4201,"failed":0} SS ------------------------------ [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:44:57.947: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-6401.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-6401.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6401.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-6401.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-6401.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6401.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe /etc/hosts STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:45:02.043: INFO: DNS probes using dns-6401/dns-test-a5c7d5dc-8fbf-40e7-b4ba-441ed0e6d020 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:45:02.068: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-6401" for this suite. •{"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]","total":275,"completed":233,"skipped":4203,"failed":0} SSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:45:02.092: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpa': should get the expected 'State' STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpof': should get the expected 'State' STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpn': should get the expected 'State' STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:45:35.997: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-4477" for this suite. • [SLOW TEST:33.911 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:40 when starting a container that exits /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:41 should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","total":275,"completed":234,"skipped":4215,"failed":0} SSS ------------------------------ [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:45:36.003: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should verify ResourceQuota with best effort scope. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a ResourceQuota with best effort scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a ResourceQuota with not best effort scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a best-effort pod STEP: Ensuring resource quota with best effort scope captures the pod usage STEP: Ensuring resource quota with not best effort ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage STEP: Creating a not best-effort pod STEP: Ensuring resource quota with not best effort scope captures the pod usage STEP: Ensuring resource quota with best effort scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:45:52.588: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-3079" for this suite. • [SLOW TEST:16.594 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should verify ResourceQuota with best effort scope. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":275,"completed":235,"skipped":4218,"failed":0} SSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:45:52.598: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:219 [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:45:52.690: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9947' Apr 25 00:45:52.949: INFO: stderr: "" Apr 25 00:45:52.949: INFO: stdout: "replicationcontroller/agnhost-master created\n" Apr 25 00:45:52.949: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9947' Apr 25 00:45:53.200: INFO: stderr: "" Apr 25 00:45:53.200: INFO: stdout: "service/agnhost-master created\n" STEP: Waiting for Agnhost master to start. Apr 25 00:45:54.206: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:45:54.206: INFO: Found 0 / 1 Apr 25 00:45:55.205: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:45:55.205: INFO: Found 0 / 1 Apr 25 00:45:56.206: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:45:56.206: INFO: Found 1 / 1 Apr 25 00:45:56.206: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Apr 25 00:45:56.209: INFO: Selector matched 1 pods for map[app:agnhost] Apr 25 00:45:56.209: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Apr 25 00:45:56.209: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config describe pod agnhost-master-bfzzv --namespace=kubectl-9947' Apr 25 00:45:56.342: INFO: stderr: "" Apr 25 00:45:56.342: INFO: stdout: "Name: agnhost-master-bfzzv\nNamespace: kubectl-9947\nPriority: 0\nNode: latest-worker/172.17.0.13\nStart Time: Sat, 25 Apr 2020 00:45:52 +0000\nLabels: app=agnhost\n role=master\nAnnotations: \nStatus: Running\nIP: 10.244.2.251\nIPs:\n IP: 10.244.2.251\nControlled By: ReplicationController/agnhost-master\nContainers:\n agnhost-master:\n Container ID: containerd://b7e78b676159d72ce617f7c0b1de0ed1727e6453158fa3dd7eff5a6a9371825b\n Image: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12\n Image ID: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost@sha256:1d7f0d77a6f07fd507f147a38d06a7c8269ebabd4f923bfe46d4fb8b396a520c\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Sat, 25 Apr 2020 00:45:55 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from default-token-s5hn5 (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n default-token-s5hn5:\n Type: Secret (a volume populated by a Secret)\n SecretName: default-token-s5hn5\n Optional: false\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute for 300s\n node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled default-scheduler Successfully assigned kubectl-9947/agnhost-master-bfzzv to latest-worker\n Normal Pulled 2s kubelet, latest-worker Container image \"us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12\" already present on machine\n Normal Created 2s kubelet, latest-worker Created container agnhost-master\n Normal Started 1s kubelet, latest-worker Started container agnhost-master\n" Apr 25 00:45:56.342: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config describe rc agnhost-master --namespace=kubectl-9947' Apr 25 00:45:56.458: INFO: stderr: "" Apr 25 00:45:56.458: INFO: stdout: "Name: agnhost-master\nNamespace: kubectl-9947\nSelector: app=agnhost,role=master\nLabels: app=agnhost\n role=master\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=master\n Containers:\n agnhost-master:\n Image: us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 4s replication-controller Created pod: agnhost-master-bfzzv\n" Apr 25 00:45:56.458: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config describe service agnhost-master --namespace=kubectl-9947' Apr 25 00:45:56.554: INFO: stderr: "" Apr 25 00:45:56.554: INFO: stdout: "Name: agnhost-master\nNamespace: kubectl-9947\nLabels: app=agnhost\n role=master\nAnnotations: \nSelector: app=agnhost,role=master\nType: ClusterIP\nIP: 10.96.142.216\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 10.244.2.251:6379\nSession Affinity: None\nEvents: \n" Apr 25 00:45:56.558: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config describe node latest-control-plane' Apr 25 00:45:56.671: INFO: stderr: "" Apr 25 00:45:56.671: INFO: stdout: "Name: latest-control-plane\nRoles: master\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=latest-control-plane\n kubernetes.io/os=linux\n node-role.kubernetes.io/master=\nAnnotations: kubeadm.alpha.kubernetes.io/cri-socket: /run/containerd/containerd.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Sun, 15 Mar 2020 18:27:32 +0000\nTaints: node-role.kubernetes.io/master:NoSchedule\nUnschedulable: false\nLease:\n HolderIdentity: latest-control-plane\n AcquireTime: \n RenewTime: Sat, 25 Apr 2020 00:45:53 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n MemoryPressure False Sat, 25 Apr 2020 00:43:39 +0000 Sun, 15 Mar 2020 18:27:32 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Sat, 25 Apr 2020 00:43:39 +0000 Sun, 15 Mar 2020 18:27:32 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Sat, 25 Apr 2020 00:43:39 +0000 Sun, 15 Mar 2020 18:27:32 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Sat, 25 Apr 2020 00:43:39 +0000 Sun, 15 Mar 2020 18:28:05 +0000 KubeletReady kubelet is posting ready status\nAddresses:\n InternalIP: 172.17.0.11\n Hostname: latest-control-plane\nCapacity:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131759892Ki\n pods: 110\nAllocatable:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131759892Ki\n pods: 110\nSystem Info:\n Machine ID: 96fd1b5d260b433d8f617f455164eb5a\n System UUID: 611bedf3-8581-4e6e-a43b-01a437bb59ad\n Boot ID: ca2aa731-f890-4956-92a1-ff8c7560d571\n Kernel Version: 4.15.0-88-generic\n OS Image: Ubuntu 19.10\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: containerd://1.3.2\n Kubelet Version: v1.17.0\n Kube-Proxy Version: v1.17.0\nPodCIDR: 10.244.0.0/24\nPodCIDRs: 10.244.0.0/24\nNon-terminated Pods: (9 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE\n --------- ---- ------------ ---------- --------------- ------------- ---\n kube-system coredns-6955765f44-f7wtl 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 40d\n kube-system coredns-6955765f44-lq4t7 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 40d\n kube-system etcd-latest-control-plane 0 (0%) 0 (0%) 0 (0%) 0 (0%) 40d\n kube-system kindnet-sx5s7 100m (0%) 100m (0%) 50Mi (0%) 50Mi (0%) 40d\n kube-system kube-apiserver-latest-control-plane 250m (1%) 0 (0%) 0 (0%) 0 (0%) 40d\n kube-system kube-controller-manager-latest-control-plane 200m (1%) 0 (0%) 0 (0%) 0 (0%) 40d\n kube-system kube-proxy-jpqvf 0 (0%) 0 (0%) 0 (0%) 0 (0%) 40d\n kube-system kube-scheduler-latest-control-plane 100m (0%) 0 (0%) 0 (0%) 0 (0%) 40d\n local-path-storage local-path-provisioner-7745554f7f-fmsmz 0 (0%) 0 (0%) 0 (0%) 0 (0%) 40d\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 850m (5%) 100m (0%)\n memory 190Mi (0%) 390Mi (0%)\n ephemeral-storage 0 (0%) 0 (0%)\n hugepages-1Gi 0 (0%) 0 (0%)\n hugepages-2Mi 0 (0%) 0 (0%)\nEvents: \n" Apr 25 00:45:56.671: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config describe namespace kubectl-9947' Apr 25 00:45:56.778: INFO: stderr: "" Apr 25 00:45:56.778: INFO: stdout: "Name: kubectl-9947\nLabels: e2e-framework=kubectl\n e2e-run=89e93e23-fc35-4cb8-a25e-45b45f06c15b\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:45:56.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9947" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance]","total":275,"completed":236,"skipped":4225,"failed":0} S ------------------------------ [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:45:56.815: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the rc STEP: delete the rc STEP: wait for all pods to be garbage collected STEP: Gathering metrics W0425 00:46:06.941928 8 metrics_grabber.go:84] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 25 00:46:06.941: INFO: For apiserver_request_total: For apiserver_request_latency_seconds: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:06.942: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-6417" for this suite. • [SLOW TEST:10.135 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","total":275,"completed":237,"skipped":4226,"failed":0} SSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:06.950: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test override command Apr 25 00:46:07.136: INFO: Waiting up to 5m0s for pod "client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f" in namespace "containers-7875" to be "Succeeded or Failed" Apr 25 00:46:07.172: INFO: Pod "client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f": Phase="Pending", Reason="", readiness=false. Elapsed: 36.435878ms Apr 25 00:46:09.175: INFO: Pod "client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039276913s Apr 25 00:46:11.179: INFO: Pod "client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04340007s STEP: Saw pod success Apr 25 00:46:11.179: INFO: Pod "client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f" satisfied condition "Succeeded or Failed" Apr 25 00:46:11.182: INFO: Trying to get logs from node latest-worker pod client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f container test-container: STEP: delete the pod Apr 25 00:46:11.240: INFO: Waiting for pod client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f to disappear Apr 25 00:46:11.245: INFO: Pod client-containers-f3aeac67-5698-4b4b-9080-f047d0648e7f no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:11.245: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-7875" for this suite. •{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":275,"completed":238,"skipped":4238,"failed":0} SSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:11.254: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD preserving unknown fields in an embedded object [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:46:11.314: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Apr 25 00:46:14.224: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4309 create -f -' Apr 25 00:46:17.016: INFO: stderr: "" Apr 25 00:46:17.016: INFO: stdout: "e2e-test-crd-publish-openapi-2735-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" Apr 25 00:46:17.016: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4309 delete e2e-test-crd-publish-openapi-2735-crds test-cr' Apr 25 00:46:17.130: INFO: stderr: "" Apr 25 00:46:17.130: INFO: stdout: "e2e-test-crd-publish-openapi-2735-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" Apr 25 00:46:17.130: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4309 apply -f -' Apr 25 00:46:17.433: INFO: stderr: "" Apr 25 00:46:17.433: INFO: stdout: "e2e-test-crd-publish-openapi-2735-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" Apr 25 00:46:17.433: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4309 delete e2e-test-crd-publish-openapi-2735-crds test-cr' Apr 25 00:46:17.523: INFO: stderr: "" Apr 25 00:46:17.523: INFO: stdout: "e2e-test-crd-publish-openapi-2735-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR Apr 25 00:46:17.524: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-2735-crds' Apr 25 00:46:17.748: INFO: stderr: "" Apr 25 00:46:17.748: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-2735-crd\nVERSION: crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Waldo\n\n status\t\n Status of Waldo\n\n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:19.662: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-4309" for this suite. • [SLOW TEST:8.415 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD preserving unknown fields in an embedded object [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":275,"completed":239,"skipped":4245,"failed":0} SSSSS ------------------------------ [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:19.670: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:46:19.797: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"bea0fb97-2168-4da6-95ff-f2f916701876", Controller:(*bool)(0xc0044a8c6a), BlockOwnerDeletion:(*bool)(0xc0044a8c6b)}} Apr 25 00:46:19.824: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"fc56defe-ae6c-4555-9846-657e1b6b68c0", Controller:(*bool)(0xc00444183a), BlockOwnerDeletion:(*bool)(0xc00444183b)}} Apr 25 00:46:19.848: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"ea1a2c2b-f40e-45e1-b232-188463e9f9a1", Controller:(*bool)(0xc00447230a), BlockOwnerDeletion:(*bool)(0xc00447230b)}} [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:24.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-1652" for this suite. • [SLOW TEST:5.230 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","total":275,"completed":240,"skipped":4250,"failed":0} SSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:24.900: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 25 00:46:28.057: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:28.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-9940" for this suite. •{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":275,"completed":241,"skipped":4254,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:28.089: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0666 on node default medium Apr 25 00:46:28.174: INFO: Waiting up to 5m0s for pod "pod-259feb10-5c1a-44c2-a87a-859b00b7bb57" in namespace "emptydir-3500" to be "Succeeded or Failed" Apr 25 00:46:28.203: INFO: Pod "pod-259feb10-5c1a-44c2-a87a-859b00b7bb57": Phase="Pending", Reason="", readiness=false. Elapsed: 29.275833ms Apr 25 00:46:30.206: INFO: Pod "pod-259feb10-5c1a-44c2-a87a-859b00b7bb57": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032460744s Apr 25 00:46:32.209: INFO: Pod "pod-259feb10-5c1a-44c2-a87a-859b00b7bb57": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035920972s STEP: Saw pod success Apr 25 00:46:32.210: INFO: Pod "pod-259feb10-5c1a-44c2-a87a-859b00b7bb57" satisfied condition "Succeeded or Failed" Apr 25 00:46:32.212: INFO: Trying to get logs from node latest-worker pod pod-259feb10-5c1a-44c2-a87a-859b00b7bb57 container test-container: STEP: delete the pod Apr 25 00:46:32.229: INFO: Waiting for pod pod-259feb10-5c1a-44c2-a87a-859b00b7bb57 to disappear Apr 25 00:46:32.234: INFO: Pod pod-259feb10-5c1a-44c2-a87a-859b00b7bb57 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:32.234: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3500" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":242,"skipped":4270,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:32.240: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: getting the auto-created API token STEP: reading a file in the container Apr 25 00:46:36.807: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1827 pod-service-account-3dff0f63-3a30-4a7b-8739-1fb0ce576986 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' STEP: reading a file in the container Apr 25 00:46:37.052: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1827 pod-service-account-3dff0f63-3a30-4a7b-8739-1fb0ce576986 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' STEP: reading a file in the container Apr 25 00:46:37.262: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1827 pod-service-account-3dff0f63-3a30-4a7b-8739-1fb0ce576986 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' [AfterEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:37.467: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-1827" for this suite. • [SLOW TEST:5.232 seconds] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods [Conformance]","total":275,"completed":243,"skipped":4294,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:37.473: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:178 [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Apr 25 00:46:42.082: INFO: Successfully updated pod "pod-update-activedeadlineseconds-d72902ba-ddbc-49f9-886f-50bcdbd2b454" Apr 25 00:46:42.082: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-d72902ba-ddbc-49f9-886f-50bcdbd2b454" in namespace "pods-2387" to be "terminated due to deadline exceeded" Apr 25 00:46:42.088: INFO: Pod "pod-update-activedeadlineseconds-d72902ba-ddbc-49f9-886f-50bcdbd2b454": Phase="Running", Reason="", readiness=true. Elapsed: 6.555838ms Apr 25 00:46:44.091: INFO: Pod "pod-update-activedeadlineseconds-d72902ba-ddbc-49f9-886f-50bcdbd2b454": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.009501264s Apr 25 00:46:44.091: INFO: Pod "pod-update-activedeadlineseconds-d72902ba-ddbc-49f9-886f-50bcdbd2b454" satisfied condition "terminated due to deadline exceeded" [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:44.091: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-2387" for this suite. • [SLOW TEST:6.627 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","total":275,"completed":244,"skipped":4321,"failed":0} S ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:44.100: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a replica set. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ReplicaSet STEP: Ensuring resource quota status captures replicaset creation STEP: Deleting a ReplicaSet STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:55.244: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-8607" for this suite. • [SLOW TEST:11.151 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a replica set. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","total":275,"completed":245,"skipped":4322,"failed":0} SSSSSS ------------------------------ [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:55.251: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-4df92c82-d009-4791-8fa9-b806b1f7d0a0 STEP: Creating a pod to test consume secrets Apr 25 00:46:55.308: INFO: Waiting up to 5m0s for pod "pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf" in namespace "secrets-3403" to be "Succeeded or Failed" Apr 25 00:46:55.312: INFO: Pod "pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf": Phase="Pending", Reason="", readiness=false. Elapsed: 3.59238ms Apr 25 00:46:57.316: INFO: Pod "pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008307342s Apr 25 00:46:59.322: INFO: Pod "pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013450492s STEP: Saw pod success Apr 25 00:46:59.322: INFO: Pod "pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf" satisfied condition "Succeeded or Failed" Apr 25 00:46:59.325: INFO: Trying to get logs from node latest-worker2 pod pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf container secret-volume-test: STEP: delete the pod Apr 25 00:46:59.372: INFO: Waiting for pod pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf to disappear Apr 25 00:46:59.384: INFO: Pod pod-secrets-495bb6a4-0112-4777-9d25-7295a75cbedf no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:46:59.384: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3403" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":275,"completed":246,"skipped":4328,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:46:59.393: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD without validation schema [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:46:59.431: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Apr 25 00:47:02.362: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6244 create -f -' Apr 25 00:47:05.254: INFO: stderr: "" Apr 25 00:47:05.254: INFO: stdout: "e2e-test-crd-publish-openapi-3212-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" Apr 25 00:47:05.254: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6244 delete e2e-test-crd-publish-openapi-3212-crds test-cr' Apr 25 00:47:05.366: INFO: stderr: "" Apr 25 00:47:05.366: INFO: stdout: "e2e-test-crd-publish-openapi-3212-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" Apr 25 00:47:05.366: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6244 apply -f -' Apr 25 00:47:05.614: INFO: stderr: "" Apr 25 00:47:05.614: INFO: stdout: "e2e-test-crd-publish-openapi-3212-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" Apr 25 00:47:05.614: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6244 delete e2e-test-crd-publish-openapi-3212-crds test-cr' Apr 25 00:47:05.722: INFO: stderr: "" Apr 25 00:47:05.722: INFO: stdout: "e2e-test-crd-publish-openapi-3212-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR without validation schema Apr 25 00:47:05.722: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:32771 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-3212-crds' Apr 25 00:47:05.996: INFO: stderr: "" Apr 25 00:47:05.996: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-3212-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:47:08.894: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-6244" for this suite. • [SLOW TEST:9.508 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD without validation schema [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":275,"completed":247,"skipped":4343,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:47:08.902: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:47:08.984: INFO: Waiting up to 5m0s for pod "busybox-user-65534-87b680ad-9e2f-47d7-b393-0500c665999e" in namespace "security-context-test-2777" to be "Succeeded or Failed" Apr 25 00:47:08.987: INFO: Pod "busybox-user-65534-87b680ad-9e2f-47d7-b393-0500c665999e": Phase="Pending", Reason="", readiness=false. Elapsed: 3.316593ms Apr 25 00:47:10.990: INFO: Pod "busybox-user-65534-87b680ad-9e2f-47d7-b393-0500c665999e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006180202s Apr 25 00:47:12.994: INFO: Pod "busybox-user-65534-87b680ad-9e2f-47d7-b393-0500c665999e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010128402s Apr 25 00:47:12.994: INFO: Pod "busybox-user-65534-87b680ad-9e2f-47d7-b393-0500c665999e" satisfied condition "Succeeded or Failed" [AfterEach] [k8s.io] Security Context /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:47:12.994: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-2777" for this suite. •{"msg":"PASSED [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":248,"skipped":4362,"failed":0} SSSSSSSS ------------------------------ [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:47:13.001: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:74 [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:47:13.092: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) Apr 25 00:47:13.113: INFO: Pod name sample-pod: Found 0 pods out of 1 Apr 25 00:47:18.126: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Apr 25 00:47:18.126: INFO: Creating deployment "test-rolling-update-deployment" Apr 25 00:47:18.138: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has Apr 25 00:47:18.158: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created Apr 25 00:47:20.166: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected Apr 25 00:47:20.168: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372438, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372438, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372438, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372438, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-664dd8fc7f\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:47:22.173: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:68 Apr 25 00:47:22.182: INFO: Deployment "test-rolling-update-deployment": &Deployment{ObjectMeta:{test-rolling-update-deployment deployment-1947 /apis/apps/v1/namespaces/deployment-1947/deployments/test-rolling-update-deployment 5828cd32-2e5d-4197-bde0-7ff8dd82da6f 10801902 1 2020-04-25 00:47:18 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0021b7248 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-04-25 00:47:18 +0000 UTC,LastTransitionTime:2020-04-25 00:47:18 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-664dd8fc7f" has successfully progressed.,LastUpdateTime:2020-04-25 00:47:21 +0000 UTC,LastTransitionTime:2020-04-25 00:47:18 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Apr 25 00:47:22.185: INFO: New ReplicaSet "test-rolling-update-deployment-664dd8fc7f" of Deployment "test-rolling-update-deployment": &ReplicaSet{ObjectMeta:{test-rolling-update-deployment-664dd8fc7f deployment-1947 /apis/apps/v1/namespaces/deployment-1947/replicasets/test-rolling-update-deployment-664dd8fc7f cab17e06-65a9-4ecb-80a7-940309c30411 10801890 1 2020-04-25 00:47:18 +0000 UTC map[name:sample-pod pod-template-hash:664dd8fc7f] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 5828cd32-2e5d-4197-bde0-7ff8dd82da6f 0xc00346c007 0xc00346c008}] [] []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 664dd8fc7f,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:664dd8fc7f] map[] [] [] []} {[] [] [{agnhost us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00346c108 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:47:22.185: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": Apr 25 00:47:22.185: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-1947 /apis/apps/v1/namespaces/deployment-1947/replicasets/test-rolling-update-controller aa6b41f7-1892-419c-a625-d1e484220631 10801901 2 2020-04-25 00:47:13 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 5828cd32-2e5d-4197-bde0-7ff8dd82da6f 0xc004051f27 0xc004051f28}] [] []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc004051f88 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:47:22.188: INFO: Pod "test-rolling-update-deployment-664dd8fc7f-kgvs7" is available: &Pod{ObjectMeta:{test-rolling-update-deployment-664dd8fc7f-kgvs7 test-rolling-update-deployment-664dd8fc7f- deployment-1947 /api/v1/namespaces/deployment-1947/pods/test-rolling-update-deployment-664dd8fc7f-kgvs7 151a7c7d-b695-455c-8d5f-fd82ff5ad099 10801889 0 2020-04-25 00:47:18 +0000 UTC map[name:sample-pod pod-template-hash:664dd8fc7f] map[] [{apps/v1 ReplicaSet test-rolling-update-deployment-664dd8fc7f cab17e06-65a9-4ecb-80a7-940309c30411 0xc0024882e7 0xc0024882e8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-8q4s9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-8q4s9,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-8q4s9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:47:18 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:47:21 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:47:21 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:47:18 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:10.244.1.33,StartTime:2020-04-25 00:47:18 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-04-25 00:47:20 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12,ImageID:us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost@sha256:1d7f0d77a6f07fd507f147a38d06a7c8269ebabd4f923bfe46d4fb8b396a520c,ContainerID:containerd://a47c9841e5122a1443a8a7d677834076b65a3d1722485ccb9f8da94387f6148d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.33,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:47:22.188: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-1947" for this suite. • [SLOW TEST:9.195 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":275,"completed":249,"skipped":4370,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:47:22.196: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating pod liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 in namespace container-probe-5768 Apr 25 00:47:26.292: INFO: Started pod liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 in namespace container-probe-5768 STEP: checking the pod's current state and verifying that restartCount is present Apr 25 00:47:26.296: INFO: Initial restart count of pod liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 is 0 Apr 25 00:47:38.322: INFO: Restart count of pod container-probe-5768/liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 is now 1 (12.026596s elapsed) Apr 25 00:47:58.363: INFO: Restart count of pod container-probe-5768/liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 is now 2 (32.067661789s elapsed) Apr 25 00:48:18.407: INFO: Restart count of pod container-probe-5768/liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 is now 3 (52.110982656s elapsed) Apr 25 00:48:38.448: INFO: Restart count of pod container-probe-5768/liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 is now 4 (1m12.152481546s elapsed) Apr 25 00:49:52.650: INFO: Restart count of pod container-probe-5768/liveness-5da32cbe-aaaa-42ab-a302-b5f4baa127b9 is now 5 (2m26.354521544s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:49:52.674: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-5768" for this suite. • [SLOW TEST:150.508 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":275,"completed":250,"skipped":4388,"failed":0} SSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:49:52.704: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:49:52.777: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54" in namespace "projected-7717" to be "Succeeded or Failed" Apr 25 00:49:52.781: INFO: Pod "downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54": Phase="Pending", Reason="", readiness=false. Elapsed: 3.965091ms Apr 25 00:49:54.786: INFO: Pod "downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008221553s Apr 25 00:49:56.818: INFO: Pod "downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.041145645s STEP: Saw pod success Apr 25 00:49:56.819: INFO: Pod "downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54" satisfied condition "Succeeded or Failed" Apr 25 00:49:56.821: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54 container client-container: STEP: delete the pod Apr 25 00:49:56.850: INFO: Waiting for pod downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54 to disappear Apr 25 00:49:56.854: INFO: Pod downwardapi-volume-a3449494-0401-440d-90e6-ddb6dd9eac54 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:49:56.854: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7717" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":275,"completed":251,"skipped":4397,"failed":0} SS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:49:56.859: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:49:57.656: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:49:59.667: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372597, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372597, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372597, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372597, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:50:02.704: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] patching/updating a validating webhook should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a validating webhook configuration STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Updating a validating webhook configuration's rules to not include the create operation STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Patching a validating webhook configuration's rules to include the create operation STEP: Creating a configMap that does not comply to the validation webhook rules [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:02.808: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-9453" for this suite. STEP: Destroying namespace "webhook-9453-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:6.088 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 patching/updating a validating webhook should work [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","total":275,"completed":252,"skipped":4399,"failed":0} SS ------------------------------ [k8s.io] Pods should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:02.948: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:178 [It] should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating pod Apr 25 00:50:07.013: INFO: Pod pod-hostip-c3c7ae74-8b5c-4b7e-bbd1-270a7a019c72 has hostIP: 172.17.0.12 [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:07.013: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-7382" for this suite. •{"msg":"PASSED [k8s.io] Pods should get a host IP [NodeConformance] [Conformance]","total":275,"completed":253,"skipped":4401,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:07.023: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a job STEP: Ensuring job reaches completions [AfterEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:21.238: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-8151" for this suite. • [SLOW TEST:14.223 seconds] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":275,"completed":254,"skipped":4438,"failed":0} SS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:21.246: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:50:22.238: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:50:24.249: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372622, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372622, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372622, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372622, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:50:27.276: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate configmap [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering the mutating configmap webhook via the AdmissionRegistration API STEP: create a configmap that should be updated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:27.363: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-9949" for this suite. STEP: Destroying namespace "webhook-9949-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:6.202 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate configmap [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":275,"completed":255,"skipped":4440,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:27.449: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:50:28.244: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:50:30.257: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372628, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372628, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372628, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372628, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:50:33.289: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:50:33.293: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6858-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource that should be mutated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:34.389: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-1555" for this suite. STEP: Destroying namespace "webhook-1555-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:7.047 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":275,"completed":256,"skipped":4461,"failed":0} SSSSSSS ------------------------------ [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:34.496: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:50:34.623: INFO: Creating ReplicaSet my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5 Apr 25 00:50:34.663: INFO: Pod name my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5: Found 0 pods out of 1 Apr 25 00:50:39.681: INFO: Pod name my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5: Found 1 pods out of 1 Apr 25 00:50:39.681: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5" is running Apr 25 00:50:39.684: INFO: Pod "my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5-vdmtr" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-25 00:50:34 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-25 00:50:37 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-25 00:50:37 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-25 00:50:34 +0000 UTC Reason: Message:}]) Apr 25 00:50:39.684: INFO: Trying to dial the pod Apr 25 00:50:44.696: INFO: Controller my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5: Got expected result from replica 1 [my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5-vdmtr]: "my-hostname-basic-87dfa9c5-3389-4ea1-9b28-5328a35361f5-vdmtr", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:44.696: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-4851" for this suite. • [SLOW TEST:10.207 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance]","total":275,"completed":257,"skipped":4468,"failed":0} SSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:44.703: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test override all Apr 25 00:50:44.773: INFO: Waiting up to 5m0s for pod "client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7" in namespace "containers-5164" to be "Succeeded or Failed" Apr 25 00:50:44.777: INFO: Pod "client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.175633ms Apr 25 00:50:46.807: INFO: Pod "client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033466155s Apr 25 00:50:48.811: INFO: Pod "client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037568733s STEP: Saw pod success Apr 25 00:50:48.811: INFO: Pod "client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7" satisfied condition "Succeeded or Failed" Apr 25 00:50:48.814: INFO: Trying to get logs from node latest-worker2 pod client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7 container test-container: STEP: delete the pod Apr 25 00:50:48.839: INFO: Waiting for pod client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7 to disappear Apr 25 00:50:48.896: INFO: Pod client-containers-83f6b17c-3c09-4611-8e2c-947d9a6978c7 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:48.896: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-5164" for this suite. •{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","total":275,"completed":258,"skipped":4476,"failed":0} SSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:48.905: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating secret with name secret-test-map-c631d9dd-4fbf-4b91-89bc-dabfc3fcafda STEP: Creating a pod to test consume secrets Apr 25 00:50:48.989: INFO: Waiting up to 5m0s for pod "pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1" in namespace "secrets-4519" to be "Succeeded or Failed" Apr 25 00:50:48.993: INFO: Pod "pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1": Phase="Pending", Reason="", readiness=false. Elapsed: 3.803876ms Apr 25 00:50:50.997: INFO: Pod "pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008206452s Apr 25 00:50:53.001: INFO: Pod "pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012108175s STEP: Saw pod success Apr 25 00:50:53.001: INFO: Pod "pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1" satisfied condition "Succeeded or Failed" Apr 25 00:50:53.003: INFO: Trying to get logs from node latest-worker2 pod pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1 container secret-volume-test: STEP: delete the pod Apr 25 00:50:53.018: INFO: Waiting for pod pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1 to disappear Apr 25 00:50:53.023: INFO: Pod pod-secrets-c71a78f4-172d-400a-bc33-b9a2b08296e1 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:53.023: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-4519" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":259,"skipped":4481,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:53.048: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for pods for Hostname [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-3907.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-3907.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3907.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-3907.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-3907.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3907.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 25 00:50:59.224: INFO: DNS probes using dns-3907/dns-test-bd5a25e1-9151-4a17-9e74-8e6fae1e9e85 succeeded STEP: deleting the pod STEP: deleting the test headless service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:50:59.325: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-3907" for this suite. • [SLOW TEST:6.288 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for pods for Hostname [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance]","total":275,"completed":260,"skipped":4496,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:50:59.336: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward api env vars Apr 25 00:50:59.445: INFO: Waiting up to 5m0s for pod "downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3" in namespace "downward-api-5066" to be "Succeeded or Failed" Apr 25 00:50:59.449: INFO: Pod "downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3": Phase="Pending", Reason="", readiness=false. Elapsed: 3.261917ms Apr 25 00:51:01.461: INFO: Pod "downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015836396s Apr 25 00:51:03.466: INFO: Pod "downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020193931s STEP: Saw pod success Apr 25 00:51:03.466: INFO: Pod "downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3" satisfied condition "Succeeded or Failed" Apr 25 00:51:03.468: INFO: Trying to get logs from node latest-worker2 pod downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3 container dapi-container: STEP: delete the pod Apr 25 00:51:03.486: INFO: Waiting for pod downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3 to disappear Apr 25 00:51:03.490: INFO: Pod downward-api-465e3fd9-586d-4857-a066-c6b4ea2d86b3 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:51:03.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5066" for this suite. •{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":275,"completed":261,"skipped":4515,"failed":0} SS ------------------------------ [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:51:03.498: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should verify ResourceQuota with terminating scopes. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a ResourceQuota with terminating scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a ResourceQuota with not terminating scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a long running pod STEP: Ensuring resource quota with not terminating scope captures the pod usage STEP: Ensuring resource quota with terminating scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage STEP: Creating a terminating pod STEP: Ensuring resource quota with terminating scope captures the pod usage STEP: Ensuring resource quota with not terminating scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:51:19.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-8047" for this suite. • [SLOW TEST:16.220 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should verify ResourceQuota with terminating scopes. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","total":275,"completed":262,"skipped":4517,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:51:19.718: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test downward API volume plugin Apr 25 00:51:19.925: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71" in namespace "downward-api-6729" to be "Succeeded or Failed" Apr 25 00:51:19.958: INFO: Pod "downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71": Phase="Pending", Reason="", readiness=false. Elapsed: 33.219975ms Apr 25 00:51:21.989: INFO: Pod "downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063425737s Apr 25 00:51:23.992: INFO: Pod "downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71": Phase="Running", Reason="", readiness=true. Elapsed: 4.066724378s Apr 25 00:51:25.995: INFO: Pod "downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.069840908s STEP: Saw pod success Apr 25 00:51:25.995: INFO: Pod "downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71" satisfied condition "Succeeded or Failed" Apr 25 00:51:25.998: INFO: Trying to get logs from node latest-worker2 pod downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71 container client-container: STEP: delete the pod Apr 25 00:51:26.042: INFO: Waiting for pod downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71 to disappear Apr 25 00:51:26.054: INFO: Pod downwardapi-volume-a162c5b3-fcf3-47e9-9202-5a5627cafc71 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:51:26.054: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6729" for this suite. • [SLOW TEST:6.355 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":263,"skipped":4531,"failed":0} SSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:51:26.073: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:51:26.994: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:51:29.025: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372687, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372687, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372687, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372686, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:51:32.043: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource with different stored version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:51:32.052: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-9967-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource while v1 is storage version STEP: Patching Custom Resource Definition to set v2 as storage STEP: Patching the custom resource while v2 is storage version [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:51:33.274: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-5817" for this suite. STEP: Destroying namespace "webhook-5817-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:7.292 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource with different stored version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":275,"completed":264,"skipped":4540,"failed":0} SSSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:51:33.365: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] should include custom resource definition resources in discovery documents [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: fetching the /apis discovery document STEP: finding the apiextensions.k8s.io API group in the /apis discovery document STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document STEP: fetching the /apis/apiextensions.k8s.io discovery document STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:51:33.404: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-3010" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","total":275,"completed":265,"skipped":4545,"failed":0} SSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:51:33.418: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:91 Apr 25 00:51:33.448: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 25 00:51:33.481: INFO: Waiting for terminating namespaces to be deleted... Apr 25 00:51:33.483: INFO: Logging pods the kubelet thinks is on node latest-worker before test Apr 25 00:51:33.502: INFO: kindnet-vnjgh from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:51:33.502: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:51:33.502: INFO: kube-proxy-s9v6p from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:51:33.503: INFO: Container kube-proxy ready: true, restart count 0 Apr 25 00:51:33.503: INFO: Logging pods the kubelet thinks is on node latest-worker2 before test Apr 25 00:51:33.507: INFO: sample-webhook-deployment-6cc9cc9dc-gsw56 from webhook-5817 started at 2020-04-25 00:51:27 +0000 UTC (1 container statuses recorded) Apr 25 00:51:33.507: INFO: Container sample-webhook ready: true, restart count 0 Apr 25 00:51:33.507: INFO: kindnet-zq6gp from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:51:33.507: INFO: Container kindnet-cni ready: true, restart count 0 Apr 25 00:51:33.507: INFO: kube-proxy-c5xlk from kube-system started at 2020-03-15 18:28:07 +0000 UTC (1 container statuses recorded) Apr 25 00:51:33.507: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-baac8018-1870-4f13-97b6-8951eb47adde 90 STEP: Trying to create a pod(pod1) with hostport 54321 and hostIP 127.0.0.1 and expect scheduled STEP: Trying to create another pod(pod2) with hostport 54321 but hostIP 127.0.0.2 on the node which pod1 resides and expect scheduled STEP: Trying to create a third pod(pod3) with hostport 54321, hostIP 127.0.0.2 but use UDP protocol on the node which pod2 resides STEP: removing the label kubernetes.io/e2e-baac8018-1870-4f13-97b6-8951eb47adde off the node latest-worker STEP: verifying the node doesn't have the label kubernetes.io/e2e-baac8018-1870-4f13-97b6-8951eb47adde [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:51:49.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-1191" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:82 • [SLOW TEST:16.405 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]","total":275,"completed":266,"skipped":4553,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:51:49.823: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:51:49.905: INFO: Creating daemon "daemon-set" with a node selector STEP: Initially, daemon pods should not be running on any nodes. Apr 25 00:51:49.910: INFO: Number of nodes with available pods: 0 Apr 25 00:51:49.910: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Change node label to blue, check that daemon pod is launched. Apr 25 00:51:49.947: INFO: Number of nodes with available pods: 0 Apr 25 00:51:49.947: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:50.952: INFO: Number of nodes with available pods: 0 Apr 25 00:51:50.952: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:51.954: INFO: Number of nodes with available pods: 0 Apr 25 00:51:51.954: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:52.952: INFO: Number of nodes with available pods: 1 Apr 25 00:51:52.952: INFO: Number of running nodes: 1, number of available pods: 1 STEP: Update the node label to green, and wait for daemons to be unscheduled Apr 25 00:51:52.982: INFO: Number of nodes with available pods: 1 Apr 25 00:51:52.982: INFO: Number of running nodes: 0, number of available pods: 1 Apr 25 00:51:53.987: INFO: Number of nodes with available pods: 0 Apr 25 00:51:53.987: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate Apr 25 00:51:53.995: INFO: Number of nodes with available pods: 0 Apr 25 00:51:53.995: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:55.059: INFO: Number of nodes with available pods: 0 Apr 25 00:51:55.059: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:55.999: INFO: Number of nodes with available pods: 0 Apr 25 00:51:55.999: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:57.065: INFO: Number of nodes with available pods: 0 Apr 25 00:51:57.065: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:58.000: INFO: Number of nodes with available pods: 0 Apr 25 00:51:58.000: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:51:58.999: INFO: Number of nodes with available pods: 0 Apr 25 00:51:58.999: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:00.000: INFO: Number of nodes with available pods: 0 Apr 25 00:52:00.000: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:01.000: INFO: Number of nodes with available pods: 0 Apr 25 00:52:01.000: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:01.999: INFO: Number of nodes with available pods: 0 Apr 25 00:52:01.999: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:02.999: INFO: Number of nodes with available pods: 0 Apr 25 00:52:02.999: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:04.000: INFO: Number of nodes with available pods: 0 Apr 25 00:52:04.000: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:05.009: INFO: Number of nodes with available pods: 0 Apr 25 00:52:05.009: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:06.000: INFO: Number of nodes with available pods: 1 Apr 25 00:52:06.000: INFO: Number of running nodes: 1, number of available pods: 1 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-3912, will wait for the garbage collector to delete the pods Apr 25 00:52:06.066: INFO: Deleting DaemonSet.extensions daemon-set took: 6.681585ms Apr 25 00:52:06.366: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.29882ms Apr 25 00:52:12.777: INFO: Number of nodes with available pods: 0 Apr 25 00:52:12.777: INFO: Number of running nodes: 0, number of available pods: 0 Apr 25 00:52:12.780: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-3912/daemonsets","resourceVersion":"10803623"},"items":null} Apr 25 00:52:12.782: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-3912/pods","resourceVersion":"10803623"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:52:12.811: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-3912" for this suite. • [SLOW TEST:22.995 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","total":275,"completed":267,"skipped":4576,"failed":0} SSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:52:12.818: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Apr 25 00:52:12.930: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:52:12.935: INFO: Number of nodes with available pods: 0 Apr 25 00:52:12.935: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:13.943: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:52:13.946: INFO: Number of nodes with available pods: 0 Apr 25 00:52:13.946: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:14.940: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:52:14.943: INFO: Number of nodes with available pods: 0 Apr 25 00:52:14.943: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:15.940: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:52:15.943: INFO: Number of nodes with available pods: 0 Apr 25 00:52:15.943: INFO: Node latest-worker is running more than one daemon pod Apr 25 00:52:16.940: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:52:16.944: INFO: Number of nodes with available pods: 2 Apr 25 00:52:16.944: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. Apr 25 00:52:16.962: INFO: DaemonSet pods can't tolerate node latest-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 25 00:52:17.006: INFO: Number of nodes with available pods: 2 Apr 25 00:52:17.006: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Wait for the failed daemon pod to be completely deleted. [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5471, will wait for the garbage collector to delete the pods Apr 25 00:52:18.373: INFO: Deleting DaemonSet.extensions daemon-set took: 63.956624ms Apr 25 00:52:18.473: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.265513ms Apr 25 00:52:32.777: INFO: Number of nodes with available pods: 0 Apr 25 00:52:32.777: INFO: Number of running nodes: 0, number of available pods: 0 Apr 25 00:52:32.780: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5471/daemonsets","resourceVersion":"10803758"},"items":null} Apr 25 00:52:32.783: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5471/pods","resourceVersion":"10803758"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:52:32.794: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-5471" for this suite. • [SLOW TEST:19.984 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","total":275,"completed":268,"skipped":4585,"failed":0} [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:52:32.802: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a secret. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Discovering how many secrets are in namespace by default STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Secret STEP: Ensuring resource quota status captures secret creation STEP: Deleting a secret STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:52:49.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-7077" for this suite. • [SLOW TEST:17.099 seconds] [sig-api-machinery] ResourceQuota /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a secret. [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","total":275,"completed":269,"skipped":4585,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:52:49.902: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:74 [It] RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:52:49.956: INFO: Creating deployment "test-recreate-deployment" Apr 25 00:52:49.967: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 Apr 25 00:52:50.024: INFO: deployment "test-recreate-deployment" doesn't have the required revision set Apr 25 00:52:52.031: INFO: Waiting deployment "test-recreate-deployment" to complete Apr 25 00:52:52.034: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372770, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372770, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372770, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372769, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-846c7dd955\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 25 00:52:54.038: INFO: Triggering a new rollout for deployment "test-recreate-deployment" Apr 25 00:52:54.045: INFO: Updating deployment test-recreate-deployment Apr 25 00:52:54.045: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:68 Apr 25 00:52:54.500: INFO: Deployment "test-recreate-deployment": &Deployment{ObjectMeta:{test-recreate-deployment deployment-7760 /apis/apps/v1/namespaces/deployment-7760/deployments/test-recreate-deployment 8ab968b1-53d0-4a5a-8901-edf20486cb52 10803902 2 2020-04-25 00:52:49 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0036a8668 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-04-25 00:52:54 +0000 UTC,LastTransitionTime:2020-04-25 00:52:54 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-5f94c574ff" is progressing.,LastUpdateTime:2020-04-25 00:52:54 +0000 UTC,LastTransitionTime:2020-04-25 00:52:49 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} Apr 25 00:52:54.512: INFO: New ReplicaSet "test-recreate-deployment-5f94c574ff" of Deployment "test-recreate-deployment": &ReplicaSet{ObjectMeta:{test-recreate-deployment-5f94c574ff deployment-7760 /apis/apps/v1/namespaces/deployment-7760/replicasets/test-recreate-deployment-5f94c574ff a7b9217b-4438-4290-981d-afc7cdef3e1b 10803900 1 2020-04-25 00:52:54 +0000 UTC map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 8ab968b1-53d0-4a5a-8901-edf20486cb52 0xc0036a8a67 0xc0036a8a68}] [] []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5f94c574ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0036a8ac8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:52:54.512: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": Apr 25 00:52:54.512: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-846c7dd955 deployment-7760 /apis/apps/v1/namespaces/deployment-7760/replicasets/test-recreate-deployment-846c7dd955 62c108ed-28b0-49d4-b897-5f15b128f5af 10803888 2 2020-04-25 00:52:49 +0000 UTC map[name:sample-pod-3 pod-template-hash:846c7dd955] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 8ab968b1-53d0-4a5a-8901-edf20486cb52 0xc0036a8b37 0xc0036a8b38}] [] []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 846c7dd955,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:846c7dd955] map[] [] [] []} {[] [] [{agnhost us.gcr.io/k8s-artifacts-prod/e2e-test-images/agnhost:2.12 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0036a8ba8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Apr 25 00:52:54.515: INFO: Pod "test-recreate-deployment-5f94c574ff-642dc" is not available: &Pod{ObjectMeta:{test-recreate-deployment-5f94c574ff-642dc test-recreate-deployment-5f94c574ff- deployment-7760 /api/v1/namespaces/deployment-7760/pods/test-recreate-deployment-5f94c574ff-642dc 48a25580-fe25-414a-9aa8-010f1c9eac1d 10803901 0 2020-04-25 00:52:54 +0000 UTC map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [{apps/v1 ReplicaSet test-recreate-deployment-5f94c574ff a7b9217b-4438-4290-981d-afc7cdef3e1b 0xc0036a9067 0xc0036a9068}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-7rcws,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-7rcws,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-7rcws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:latest-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:52:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:52:54 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:52:54 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-04-25 00:52:54 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.17.0.12,PodIP:,StartTime:2020-04-25 00:52:54 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:52:54.515: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-7760" for this suite. •{"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","total":275,"completed":270,"skipped":4602,"failed":0} SSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:52:54.523: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Apr 25 00:52:55.289: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Apr 25 00:52:57.324: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372775, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372775, loc:(*time.Location)(0x7b1e080)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372775, loc:(*time.Location)(0x7b1e080)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63723372775, loc:(*time.Location)(0x7b1e080)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-6cc9cc9dc\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Apr 25 00:53:00.350: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny pod and configmap creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Registering the webhook via the AdmissionRegistration API STEP: create a pod that should be denied by the webhook STEP: create a pod that causes the webhook to hang STEP: create a configmap that should be denied by the webhook STEP: create a configmap that should be admitted by the webhook STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook STEP: create a namespace that bypass the webhook STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:53:10.497: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2430" for this suite. STEP: Destroying namespace "webhook-2430-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:16.107 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny pod and configmap creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","total":275,"completed":271,"skipped":4612,"failed":0} SSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:53:10.630: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should patch a Namespace [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: creating a Namespace STEP: patching the Namespace STEP: get the Namespace and ensuring it has the label [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:53:10.829: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-4713" for this suite. STEP: Destroying namespace "nspatchtest-03a19f20-b7c7-4e1d-abf3-f23996c19f09-3405" for this suite. •{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance]","total":275,"completed":272,"skipped":4623,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Proxy version v1 should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:53:10.864: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:53:10.969: INFO: (0) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 17.318114ms) Apr 25 00:53:10.976: INFO: (1) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 7.405096ms) Apr 25 00:53:10.979: INFO: (2) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.85088ms) Apr 25 00:53:10.982: INFO: (3) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.953258ms) Apr 25 00:53:10.985: INFO: (4) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.089398ms) Apr 25 00:53:10.988: INFO: (5) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.728334ms) Apr 25 00:53:10.991: INFO: (6) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.295824ms) Apr 25 00:53:10.994: INFO: (7) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.7905ms) Apr 25 00:53:10.997: INFO: (8) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.539675ms) Apr 25 00:53:11.000: INFO: (9) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.998262ms) Apr 25 00:53:11.003: INFO: (10) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.094255ms) Apr 25 00:53:11.006: INFO: (11) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.141904ms) Apr 25 00:53:11.009: INFO: (12) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.933555ms) Apr 25 00:53:11.012: INFO: (13) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.962945ms) Apr 25 00:53:11.015: INFO: (14) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.188286ms) Apr 25 00:53:11.018: INFO: (15) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.025392ms) Apr 25 00:53:11.022: INFO: (16) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.191037ms) Apr 25 00:53:11.025: INFO: (17) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 2.944012ms) Apr 25 00:53:11.029: INFO: (18) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.917385ms) Apr 25 00:53:11.032: INFO: (19) /api/v1/nodes/latest-worker/proxy/logs/:
containers/
pods/
(200; 3.477788ms) [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:53:11.032: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-2476" for this suite. •{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node using proxy subresource [Conformance]","total":275,"completed":273,"skipped":4659,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:53:11.040: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 Apr 25 00:53:11.308: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:53:13.311: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Pending, waiting for it to be Running (with Ready = true) Apr 25 00:53:15.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:17.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:19.313: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:21.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:23.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:25.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:27.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:29.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = false) Apr 25 00:53:31.312: INFO: The status of Pod test-webserver-2d42a1ff-3c26-4b4c-ad39-f997c14ad29b is Running (Ready = true) Apr 25 00:53:31.315: INFO: Container started at 2020-04-25 00:53:13 +0000 UTC, pod became ready at 2020-04-25 00:53:30 +0000 UTC [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:53:31.315: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-3986" for this suite. • [SLOW TEST:20.283 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:698 with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 ------------------------------ {"msg":"PASSED [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","total":275,"completed":274,"skipped":4686,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:178 STEP: Creating a kubernetes client Apr 25 00:53:31.324: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:703 STEP: Creating a pod to test emptydir 0777 on tmpfs Apr 25 00:53:31.395: INFO: Waiting up to 5m0s for pod "pod-85952997-cda2-43d5-8416-212e36064370" in namespace "emptydir-1865" to be "Succeeded or Failed" Apr 25 00:53:31.399: INFO: Pod "pod-85952997-cda2-43d5-8416-212e36064370": Phase="Pending", Reason="", readiness=false. Elapsed: 3.73271ms Apr 25 00:53:33.403: INFO: Pod "pod-85952997-cda2-43d5-8416-212e36064370": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007741484s Apr 25 00:53:35.407: INFO: Pod "pod-85952997-cda2-43d5-8416-212e36064370": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012225109s STEP: Saw pod success Apr 25 00:53:35.407: INFO: Pod "pod-85952997-cda2-43d5-8416-212e36064370" satisfied condition "Succeeded or Failed" Apr 25 00:53:35.410: INFO: Trying to get logs from node latest-worker2 pod pod-85952997-cda2-43d5-8416-212e36064370 container test-container: STEP: delete the pod Apr 25 00:53:35.464: INFO: Waiting for pod pod-85952997-cda2-43d5-8416-212e36064370 to disappear Apr 25 00:53:35.485: INFO: Pod pod-85952997-cda2-43d5-8416-212e36064370 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:179 Apr 25 00:53:35.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1865" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":275,"completed":275,"skipped":4697,"failed":0} SSSSSSSSSSSSSSSSSSSSApr 25 00:53:35.491: INFO: Running AfterSuite actions on all nodes Apr 25 00:53:35.491: INFO: Running AfterSuite actions on node 1 Apr 25 00:53:35.491: INFO: Skipping dumping logs from cluster JUnit report was created: /home/opnfv/functest/results/k8s_conformance/junit_01.xml {"msg":"Test Suite completed","total":275,"completed":275,"skipped":4717,"failed":0} Ran 275 of 4992 Specs in 4577.756 seconds SUCCESS! -- 275 Passed | 0 Failed | 0 Pending | 4717 Skipped PASS