Running Suite: Kubernetes e2e suite =================================== Random Seed: 1622719549 - Will randomize all specs Will run 5771 specs Running in parallel across 10 nodes Jun 3 11:25:51.215: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.218: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable Jun 3 11:25:51.244: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready Jun 3 11:25:51.294: INFO: 21 / 21 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) Jun 3 11:25:51.294: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. Jun 3 11:25:51.294: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start Jun 3 11:25:51.309: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'create-loop-devs' (0 seconds elapsed) Jun 3 11:25:51.309: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kindnet' (0 seconds elapsed) Jun 3 11:25:51.309: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-multus-ds' (0 seconds elapsed) Jun 3 11:25:51.309: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) Jun 3 11:25:51.309: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'tune-sysctls' (0 seconds elapsed) Jun 3 11:25:51.309: INFO: e2e test version: v1.21.1 Jun 3 11:25:51.311: INFO: kube-apiserver version: v1.21.1 Jun 3 11:25:51.311: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.318: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.332: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.355: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.341: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.362: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.355: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.376: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.353: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.376: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.359: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.379: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.359: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.379: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.392: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.411: INFO: Cluster IP family: ipv4 Jun 3 11:25:51.417: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.436: INFO: Cluster IP family: ipv4 SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-instrumentation] Events /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.395: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events W0603 11:25:51.423338 34 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.423: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.426: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a collection of events [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Create set of events Jun 3 11:25:51.432: INFO: created test-event-1 Jun 3 11:25:51.434: INFO: created test-event-2 Jun 3 11:25:51.436: INFO: created test-event-3 STEP: get a list of Events with a label in the current namespace STEP: delete collection of events Jun 3 11:25:51.438: INFO: requesting DeleteCollection of events STEP: check that the list of events matches the requested quantity Jun 3 11:25:51.448: INFO: requesting list of events to confirm quantity [AfterEach] [sig-instrumentation] Events /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:25:51.450: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-1579" for this suite. •SS ------------------------------ {"msg":"PASSED [sig-instrumentation] Events should delete a collection of events [Conformance]","total":-1,"completed":1,"skipped":12,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ Jun 3 11:25:51.473: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:25:51.493: INFO: Cluster IP family: ipv4 SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.399: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir W0603 11:25:51.423892 28 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.423: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.426: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir volume type on tmpfs Jun 3 11:25:51.434: INFO: Waiting up to 5m0s for pod "pod-e0a4c651-30bc-46f5-9701-979ae5496f84" in namespace "emptydir-4090" to be "Succeeded or Failed" Jun 3 11:25:51.436: INFO: Pod "pod-e0a4c651-30bc-46f5-9701-979ae5496f84": Phase="Pending", Reason="", readiness=false. Elapsed: 1.821207ms Jun 3 11:25:53.440: INFO: Pod "pod-e0a4c651-30bc-46f5-9701-979ae5496f84": Phase="Pending", Reason="", readiness=false. Elapsed: 2.005964473s Jun 3 11:25:55.445: INFO: Pod "pod-e0a4c651-30bc-46f5-9701-979ae5496f84": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010799199s Jun 3 11:25:57.449: INFO: Pod "pod-e0a4c651-30bc-46f5-9701-979ae5496f84": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015124327s STEP: Saw pod success Jun 3 11:25:57.449: INFO: Pod "pod-e0a4c651-30bc-46f5-9701-979ae5496f84" satisfied condition "Succeeded or Failed" Jun 3 11:25:57.452: INFO: Trying to get logs from node v1.21-worker pod pod-e0a4c651-30bc-46f5-9701-979ae5496f84 container test-container: STEP: delete the pod Jun 3 11:25:57.823: INFO: Waiting for pod pod-e0a4c651-30bc-46f5-9701-979ae5496f84 to disappear Jun 3 11:25:57.826: INFO: Pod pod-e0a4c651-30bc-46f5-9701-979ae5496f84 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:25:57.826: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4090" for this suite. • [SLOW TEST:6.438 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":15,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.482: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:69 [BeforeEach] Listing PodDisruptionBudgets for all namespaces /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.505: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption-2 STEP: Waiting for a default service account to be provisioned in namespace [It] should list and delete a collection of PodDisruptionBudgets [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Waiting for the pdb to be processed STEP: Waiting for the pdb to be processed STEP: Waiting for the pdb to be processed STEP: listing a collection of PDBs across all namespaces STEP: listing a collection of PDBs in namespace disruption-8403 STEP: deleting a collection of PDBs STEP: Waiting for the PDB collection to be deleted [AfterEach] Listing PodDisruptionBudgets for all namespaces /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:25:57.832: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-2-9604" for this suite. [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:25:57.840: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-8403" for this suite. • [SLOW TEST:6.367 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Listing PodDisruptionBudgets for all namespaces /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:75 should list and delete a collection of PodDisruptionBudgets [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces should list and delete a collection of PodDisruptionBudgets [Conformance]","total":-1,"completed":2,"skipped":28,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.411: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller W0603 11:25:51.435837 26 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.435: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.438: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should adopt matching pods on creation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Given a Pod with a 'name' label pod-adoption is created Jun 3 11:25:51.448: INFO: The status of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:25:53.452: INFO: The status of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:25:55.452: INFO: The status of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:25:57.452: INFO: The status of Pod pod-adoption is Running (Ready = true) STEP: When a replication controller with a matching selector is created STEP: Then the orphan pod is adopted [AfterEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:25:58.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-5398" for this suite. • [SLOW TEST:7.067 seconds] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","total":-1,"completed":1,"skipped":30,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.461: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook W0603 11:25:51.483909 24 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.484: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.487: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:25:52.558: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:25:55.577: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource with pruning [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:25:55.581: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-5706-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource that should be mutated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:25:58.734: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2026" for this suite. STEP: Destroying namespace "webhook-2026-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:7.311 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource with pruning [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","total":-1,"completed":1,"skipped":57,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.360: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir W0603 11:25:51.389718 32 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.389: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.396: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir volume type on node default medium Jun 3 11:25:51.403: INFO: Waiting up to 5m0s for pod "pod-5ed8660a-07c6-45fe-93f6-7238cf02a793" in namespace "emptydir-3988" to be "Succeeded or Failed" Jun 3 11:25:51.405: INFO: Pod "pod-5ed8660a-07c6-45fe-93f6-7238cf02a793": Phase="Pending", Reason="", readiness=false. Elapsed: 2.268238ms Jun 3 11:25:53.410: INFO: Pod "pod-5ed8660a-07c6-45fe-93f6-7238cf02a793": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007015577s Jun 3 11:25:55.415: INFO: Pod "pod-5ed8660a-07c6-45fe-93f6-7238cf02a793": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011705539s Jun 3 11:25:57.419: INFO: Pod "pod-5ed8660a-07c6-45fe-93f6-7238cf02a793": Phase="Pending", Reason="", readiness=false. Elapsed: 6.016274529s Jun 3 11:25:59.424: INFO: Pod "pod-5ed8660a-07c6-45fe-93f6-7238cf02a793": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.020507829s STEP: Saw pod success Jun 3 11:25:59.424: INFO: Pod "pod-5ed8660a-07c6-45fe-93f6-7238cf02a793" satisfied condition "Succeeded or Failed" Jun 3 11:25:59.427: INFO: Trying to get logs from node v1.21-worker pod pod-5ed8660a-07c6-45fe-93f6-7238cf02a793 container test-container: STEP: delete the pod Jun 3 11:25:59.439: INFO: Waiting for pod pod-5ed8660a-07c6-45fe-93f6-7238cf02a793 to disappear Jun 3 11:25:59.441: INFO: Pod pod-5ed8660a-07c6-45fe-93f6-7238cf02a793 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:25:59.441: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3988" for this suite. • [SLOW TEST:8.090 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":2,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.526: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods W0603 11:25:51.828712 30 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.828: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.832: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should run through the lifecycle of Pods and PodStatus [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a Pod with a static label STEP: watching for Pod to be ready Jun 3 11:25:51.852: INFO: observed Pod pod-test in namespace pods-5202 in phase Pending with labels: map[test-pod-static:true] & conditions [] Jun 3 11:25:51.854: INFO: observed Pod pod-test in namespace pods-5202 in phase Pending with labels: map[test-pod-static:true] & conditions [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC }] Jun 3 11:25:54.763: INFO: observed Pod pod-test in namespace pods-5202 in phase Pending with labels: map[test-pod-static:true] & conditions [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC }] Jun 3 11:25:55.816: INFO: observed Pod pod-test in namespace pods-5202 in phase Pending with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC }] Jun 3 11:26:01.818: INFO: Found Pod pod-test in namespace pods-5202 in phase Running with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:55 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:55 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:25:51 +0000 UTC }] STEP: patching the Pod with a new Label and updated data Jun 3 11:26:01.833: INFO: observed event type ADDED STEP: getting the Pod and ensuring that it's patched STEP: getting the PodStatus STEP: replacing the Pod's status Ready condition to False STEP: check the Pod again to ensure its Ready conditions are False STEP: deleting the Pod via a Collection with a LabelSelector STEP: watching for the Pod to be deleted Jun 3 11:26:01.860: INFO: observed event type ADDED Jun 3 11:26:01.860: INFO: observed event type MODIFIED Jun 3 11:26:01.861: INFO: observed event type MODIFIED Jun 3 11:26:01.861: INFO: observed event type MODIFIED Jun 3 11:26:01.861: INFO: observed event type MODIFIED Jun 3 11:26:01.861: INFO: observed event type MODIFIED Jun 3 11:26:01.861: INFO: observed event type MODIFIED Jun 3 11:26:01.861: INFO: observed event type MODIFIED [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:01.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-5202" for this suite. • [SLOW TEST:10.345 seconds] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should run through the lifecycle of Pods and PodStatus [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Pods should run through the lifecycle of Pods and PodStatus [Conformance]","total":-1,"completed":1,"skipped":21,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.439: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container W0603 11:25:51.459666 17 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.459: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.463: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/init_container.go:162 [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod Jun 3 11:25:51.466: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:02.820: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-5238" for this suite. • [SLOW TEST:11.391 seconds] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","total":-1,"completed":1,"skipped":1,"failed":0} SS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.383: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods W0603 11:25:51.404781 21 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.404: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.408: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should be submitted and removed [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod STEP: setting up watch STEP: submitting the pod to kubernetes Jun 3 11:25:51.421: INFO: observed the pod list STEP: verifying the pod is in kubernetes STEP: verifying pod creation was observed STEP: deleting the pod gracefully STEP: verifying pod deletion was observed [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:05.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-2561" for this suite. • [SLOW TEST:14.410 seconds] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should be submitted and removed [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Pods should be submitted and removed [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":1,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:02.836: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should surface a failure condition on a common issue like exceeded quota [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:26:02.872: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace STEP: Creating rc "condition-test" that asks for more than the allowed pod quota STEP: Checking rc "condition-test" has the desired failure condition set STEP: Scaling down rc "condition-test" to satisfy pod quota Jun 3 11:26:05.180: INFO: Updating replication controller "condition-test" STEP: Checking rc "condition-test" has no failure condition set [AfterEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:06.386: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-8725" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":-1,"completed":2,"skipped":3,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:59.523: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:25:59.564: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa" in namespace "downward-api-335" to be "Succeeded or Failed" Jun 3 11:25:59.567: INFO: Pod "downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa": Phase="Pending", Reason="", readiness=false. Elapsed: 2.944567ms Jun 3 11:26:01.571: INFO: Pod "downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007026757s Jun 3 11:26:03.575: INFO: Pod "downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011372857s Jun 3 11:26:05.583: INFO: Pod "downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa": Phase="Pending", Reason="", readiness=false. Elapsed: 6.019727262s Jun 3 11:26:07.587: INFO: Pod "downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.023673466s STEP: Saw pod success Jun 3 11:26:07.587: INFO: Pod "downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa" satisfied condition "Succeeded or Failed" Jun 3 11:26:07.590: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa container client-container: STEP: delete the pod Jun 3 11:26:07.604: INFO: Waiting for pod downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa to disappear Jun 3 11:26:07.608: INFO: Pod downwardapi-volume-e6f7fb81-175a-4efc-a875-b3bbc98f83aa no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:07.608: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-335" for this suite. • [SLOW TEST:8.093 seconds] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":44,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:57.894: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should provide container's memory limit [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:25:57.933: INFO: Waiting up to 5m0s for pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa" in namespace "projected-2261" to be "Succeeded or Failed" Jun 3 11:25:57.937: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa": Phase="Pending", Reason="", readiness=false. Elapsed: 3.163663ms Jun 3 11:25:59.942: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008319137s Jun 3 11:26:01.946: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012403711s Jun 3 11:26:03.978: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa": Phase="Pending", Reason="", readiness=false. Elapsed: 6.04460967s Jun 3 11:26:06.178: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa": Phase="Pending", Reason="", readiness=false. Elapsed: 8.244567162s Jun 3 11:26:08.183: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa": Phase="Pending", Reason="", readiness=false. Elapsed: 10.249675469s Jun 3 11:26:10.187: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.253748709s STEP: Saw pod success Jun 3 11:26:10.187: INFO: Pod "downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa" satisfied condition "Succeeded or Failed" Jun 3 11:26:10.191: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa container client-container: STEP: delete the pod Jun 3 11:26:10.206: INFO: Waiting for pod downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa to disappear Jun 3 11:26:10.208: INFO: Pod downwardapi-volume-eac9421a-cab2-4a1f-91f9-40158b4af3fa no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:10.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2261" for this suite. • [SLOW TEST:12.324 seconds] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should provide container's memory limit [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":42,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:06.612: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-30b52e1f-edd8-4176-9138-35d797702fc0 STEP: Creating a pod to test consume secrets Jun 3 11:26:06.650: INFO: Waiting up to 5m0s for pod "pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70" in namespace "secrets-5362" to be "Succeeded or Failed" Jun 3 11:26:06.652: INFO: Pod "pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70": Phase="Pending", Reason="", readiness=false. Elapsed: 2.411921ms Jun 3 11:26:08.657: INFO: Pod "pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006925817s Jun 3 11:26:10.661: INFO: Pod "pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011580097s Jun 3 11:26:12.665: INFO: Pod "pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015673664s Jun 3 11:26:14.678: INFO: Pod "pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.027937538s STEP: Saw pod success Jun 3 11:26:14.678: INFO: Pod "pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70" satisfied condition "Succeeded or Failed" Jun 3 11:26:14.681: INFO: Trying to get logs from node v1.21-worker pod pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70 container secret-volume-test: STEP: delete the pod Jun 3 11:26:14.880: INFO: Waiting for pod pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70 to disappear Jun 3 11:26:15.485: INFO: Pod pod-secrets-b7e92529-b542-4920-8739-e21b1d030e70 no longer exists [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:15.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-5362" for this suite. • [SLOW TEST:8.972 seconds] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":-1,"completed":3,"skipped":15,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] HostPort /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:58.525: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename hostport STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] HostPort /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/hostport.go:47 [It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Trying to create a pod(pod1) with hostport 54323 and hostIP 127.0.0.1 and expect scheduled Jun 3 11:25:58.571: INFO: The status of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:00.575: INFO: The status of Pod pod1 is Running (Ready = true) STEP: Trying to create another pod(pod2) with hostport 54323 but hostIP 172.18.0.2 on the node which pod1 resides and expect scheduled Jun 3 11:26:00.583: INFO: The status of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:02.587: INFO: The status of Pod pod2 is Running (Ready = false) Jun 3 11:26:04.680: INFO: The status of Pod pod2 is Running (Ready = true) STEP: Trying to create a third pod(pod3) with hostport 54323, hostIP 172.18.0.2 but use UDP protocol on the node which pod2 resides Jun 3 11:26:04.879: INFO: The status of Pod pod3 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:06.884: INFO: The status of Pod pod3 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:08.884: INFO: The status of Pod pod3 is Running (Ready = true) Jun 3 11:26:08.891: INFO: The status of Pod e2e-host-exec is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:10.895: INFO: The status of Pod e2e-host-exec is Running (Ready = true) STEP: checking connectivity from pod e2e-host-exec to serverIP: 127.0.0.1, port: 54323 Jun 3 11:26:10.898: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 --interface 172.18.0.2 http://127.0.0.1:54323/hostname] Namespace:hostport-8719 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:26:10.898: INFO: >>> kubeConfig: /root/.kube/config STEP: checking connectivity from pod e2e-host-exec to serverIP: 172.18.0.2, port: 54323 Jun 3 11:26:11.017: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 http://172.18.0.2:54323/hostname] Namespace:hostport-8719 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:26:11.017: INFO: >>> kubeConfig: /root/.kube/config STEP: checking connectivity from pod e2e-host-exec to serverIP: 172.18.0.2, port: 54323 UDP Jun 3 11:26:11.109: INFO: ExecWithOptions {Command:[/bin/sh -c nc -vuz -w 5 172.18.0.2 54323] Namespace:hostport-8719 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:26:11.109: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-network] HostPort /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:16.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "hostport-8719" for this suite. • [SLOW TEST:18.356 seconds] [sig-network] HostPort /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] HostPort validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance]","total":-1,"completed":2,"skipped":53,"failed":0} SS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:05.844: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:26:07.780: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:26:09.792: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:11.797: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:13.798: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316367, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:26:17.381: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] listing mutating webhooks should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Listing all of the created validation webhooks STEP: Creating a configMap that should be mutated STEP: Deleting the collection of validation webhooks STEP: Creating a configMap that should not be mutated [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:17.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2577" for this suite. STEP: Destroying namespace "webhook-2577-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:11.960 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 listing mutating webhooks should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","total":-1,"completed":2,"skipped":25,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:58.819: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:126 STEP: Setting up server cert STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication STEP: Deploying the custom resource conversion webhook pod STEP: Wait for the deployment to be ready Jun 3 11:25:59.231: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set Jun 3 11:26:01.242: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-697cdbd8f4\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:03.247: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-697cdbd8f4\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:05.382: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-697cdbd8f4\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:07.245: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-697cdbd8f4\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:09.247: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316359, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-697cdbd8f4\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:26:12.255: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 [It] should be able to convert from CR v1 to CR v2 [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:26:12.258: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating a v1 custom resource STEP: v2 custom resource should be converted [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:17.887: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-webhook-8407" for this suite. [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:137 • [SLOW TEST:19.102 seconds] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to convert from CR v1 to CR v2 [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","total":-1,"completed":2,"skipped":80,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:07.716: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication Jun 3 11:26:08.183: INFO: role binding webhook-auth-reader already exists STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:26:08.197: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:26:10.207: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:12.211: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:14.212: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316368, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:26:17.883: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should deny crd creation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering the crd webhook via the AdmissionRegistration API STEP: Creating a custom resource definition that should be denied by the webhook Jun 3 11:26:17.900: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:17.922: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-5437" for this suite. STEP: Destroying namespace "webhook-5437-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:10.468 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should deny crd creation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":-1,"completed":3,"skipped":100,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.486: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath W0603 11:25:51.511066 22 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.511: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.514: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with configmap pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod pod-subpath-test-configmap-zd2p STEP: Creating a pod to test atomic-volume-subpath Jun 3 11:25:51.528: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-zd2p" in namespace "subpath-4514" to be "Succeeded or Failed" Jun 3 11:25:51.530: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Pending", Reason="", readiness=false. Elapsed: 2.416655ms Jun 3 11:25:53.534: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006260006s Jun 3 11:25:55.540: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011839084s Jun 3 11:25:57.544: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Pending", Reason="", readiness=false. Elapsed: 6.016372282s Jun 3 11:25:59.549: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Pending", Reason="", readiness=false. Elapsed: 8.020861178s Jun 3 11:26:01.553: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 10.024565557s Jun 3 11:26:03.557: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 12.029419431s Jun 3 11:26:05.583: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 14.055495639s Jun 3 11:26:07.587: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 16.059254585s Jun 3 11:26:09.592: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 18.064515043s Jun 3 11:26:11.598: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 20.07030138s Jun 3 11:26:13.606: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 22.078341416s Jun 3 11:26:16.078: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Running", Reason="", readiness=true. Elapsed: 24.549836601s Jun 3 11:26:18.088: INFO: Pod "pod-subpath-test-configmap-zd2p": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.560325091s STEP: Saw pod success Jun 3 11:26:18.088: INFO: Pod "pod-subpath-test-configmap-zd2p" satisfied condition "Succeeded or Failed" Jun 3 11:26:18.184: INFO: Trying to get logs from node v1.21-worker pod pod-subpath-test-configmap-zd2p container test-container-subpath-configmap-zd2p: STEP: delete the pod Jun 3 11:26:18.286: INFO: Waiting for pod pod-subpath-test-configmap-zd2p to disappear Jun 3 11:26:18.295: INFO: Pod pod-subpath-test-configmap-zd2p no longer exists STEP: Deleting pod pod-subpath-test-configmap-zd2p Jun 3 11:26:18.295: INFO: Deleting pod "pod-subpath-test-configmap-zd2p" in namespace "subpath-4514" [AfterEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:18.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-4514" for this suite. • [SLOW TEST:26.823 seconds] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with configmap pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]","total":-1,"completed":1,"skipped":50,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:15.639: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] custom resource defaulting for requests and from storage works [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:26:16.886: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:20.156: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-415" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance]","total":-1,"completed":4,"skipped":42,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:10.233: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] Replace and Patch tests [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:26:10.280: INFO: Pod name sample-pod: Found 0 pods out of 1 Jun 3 11:26:15.488: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running STEP: Scaling up "test-rs" replicaset Jun 3 11:26:15.586: INFO: Updating replica set "test-rs" STEP: patching the ReplicaSet Jun 3 11:26:15.784: INFO: observed ReplicaSet test-rs in namespace replicaset-7975 with ReadyReplicas 1, AvailableReplicas 1 Jun 3 11:26:16.380: INFO: observed ReplicaSet test-rs in namespace replicaset-7975 with ReadyReplicas 1, AvailableReplicas 1 Jun 3 11:26:16.881: INFO: observed ReplicaSet test-rs in namespace replicaset-7975 with ReadyReplicas 1, AvailableReplicas 1 Jun 3 11:26:16.977: INFO: observed ReplicaSet test-rs in namespace replicaset-7975 with ReadyReplicas 1, AvailableReplicas 1 Jun 3 11:26:21.825: INFO: observed ReplicaSet test-rs in namespace replicaset-7975 with ReadyReplicas 2, AvailableReplicas 2 Jun 3 11:26:22.221: INFO: observed Replicaset test-rs in namespace replicaset-7975 with ReadyReplicas 3 found true [AfterEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:22.221: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-7975" for this suite. • [SLOW TEST:11.995 seconds] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Replace and Patch tests [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet Replace and Patch tests [Conformance]","total":-1,"completed":3,"skipped":48,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:17.822: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-a03c8593-b8e3-4233-bb41-77b1192e2f5e STEP: Creating a pod to test consume secrets Jun 3 11:26:17.855: INFO: Waiting up to 5m0s for pod "pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84" in namespace "secrets-7631" to be "Succeeded or Failed" Jun 3 11:26:17.858: INFO: Pod "pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84": Phase="Pending", Reason="", readiness=false. Elapsed: 2.628471ms Jun 3 11:26:19.861: INFO: Pod "pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84": Phase="Pending", Reason="", readiness=false. Elapsed: 2.005971439s Jun 3 11:26:21.866: INFO: Pod "pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010917176s Jun 3 11:26:23.871: INFO: Pod "pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016082836s STEP: Saw pod success Jun 3 11:26:23.871: INFO: Pod "pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84" satisfied condition "Succeeded or Failed" Jun 3 11:26:23.874: INFO: Trying to get logs from node v1.21-worker pod pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84 container secret-env-test: STEP: delete the pod Jun 3 11:26:23.889: INFO: Waiting for pod pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84 to disappear Jun 3 11:26:23.891: INFO: Pod pod-secrets-6b0d7c7f-7b4a-4d38-b60f-6755b779fe84 no longer exists [AfterEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:23.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-7631" for this suite. • [SLOW TEST:6.077 seconds] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should be consumable from pods in env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","total":-1,"completed":3,"skipped":34,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:18.200: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a volume subpath [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test substitution in volume subpath Jun 3 11:26:18.321: INFO: Waiting up to 5m0s for pod "var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac" in namespace "var-expansion-7680" to be "Succeeded or Failed" Jun 3 11:26:18.324: INFO: Pod "var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.380285ms Jun 3 11:26:20.326: INFO: Pod "var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.005045428s Jun 3 11:26:22.330: INFO: Pod "var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac": Phase="Pending", Reason="", readiness=false. Elapsed: 4.009153115s Jun 3 11:26:24.334: INFO: Pod "var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01320121s STEP: Saw pod success Jun 3 11:26:24.335: INFO: Pod "var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac" satisfied condition "Succeeded or Failed" Jun 3 11:26:24.337: INFO: Trying to get logs from node v1.21-worker pod var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac container dapi-container: STEP: delete the pod Jun 3 11:26:24.351: INFO: Waiting for pod var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac to disappear Jun 3 11:26:24.353: INFO: Pod var-expansion-87e29e1d-9ff4-4f89-8de6-2a6d762bdbac no longer exists [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:24.353: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-7680" for this suite. • [SLOW TEST:6.161 seconds] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should allow substituting values in a volume subpath [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a volume subpath [Conformance]","total":-1,"completed":4,"skipped":107,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:18.364: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0777 on node default medium Jun 3 11:26:18.392: INFO: Waiting up to 5m0s for pod "pod-8b919e12-d2df-44fc-aab2-515fe56f735c" in namespace "emptydir-8639" to be "Succeeded or Failed" Jun 3 11:26:18.394: INFO: Pod "pod-8b919e12-d2df-44fc-aab2-515fe56f735c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.002477ms Jun 3 11:26:20.399: INFO: Pod "pod-8b919e12-d2df-44fc-aab2-515fe56f735c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007056288s Jun 3 11:26:22.403: INFO: Pod "pod-8b919e12-d2df-44fc-aab2-515fe56f735c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010653213s Jun 3 11:26:24.407: INFO: Pod "pod-8b919e12-d2df-44fc-aab2-515fe56f735c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015129662s STEP: Saw pod success Jun 3 11:26:24.407: INFO: Pod "pod-8b919e12-d2df-44fc-aab2-515fe56f735c" satisfied condition "Succeeded or Failed" Jun 3 11:26:24.410: INFO: Trying to get logs from node v1.21-worker pod pod-8b919e12-d2df-44fc-aab2-515fe56f735c container test-container: STEP: delete the pod Jun 3 11:26:24.421: INFO: Waiting for pod pod-8b919e12-d2df-44fc-aab2-515fe56f735c to disappear Jun 3 11:26:24.423: INFO: Pod pod-8b919e12-d2df-44fc-aab2-515fe56f735c no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:24.423: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-8639" for this suite. • [SLOW TEST:6.066 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":87,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:17.990: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:26:18.890: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created Jun 3 11:26:20.900: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:22.904: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:24.904: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316378, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:26:27.916: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] patching/updating a mutating webhook should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a mutating webhook configuration STEP: Updating a mutating webhook configuration's rules to not include the create operation STEP: Creating a configMap that should not be mutated STEP: Patching a mutating webhook configuration's rules to include the create operation STEP: Creating a configMap that should be mutated [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:27.997: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-4786" for this suite. STEP: Destroying namespace "webhook-4786-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:10.036 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 patching/updating a mutating webhook should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","total":-1,"completed":3,"skipped":126,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:22.297: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] listing custom resource definition objects works [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:26:22.328: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:28.586: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-3854" for this suite. • [SLOW TEST:6.296 seconds] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Simple CustomResourceDefinition /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:48 listing custom resource definition objects works [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance]","total":-1,"completed":4,"skipped":87,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:24.382: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46 [It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:26:24.418: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-2cd70ad8-75e0-4e2f-a3e7-373b3804a106" in namespace "security-context-test-9798" to be "Succeeded or Failed" Jun 3 11:26:24.420: INFO: Pod "busybox-readonly-false-2cd70ad8-75e0-4e2f-a3e7-373b3804a106": Phase="Pending", Reason="", readiness=false. Elapsed: 2.002759ms Jun 3 11:26:26.424: INFO: Pod "busybox-readonly-false-2cd70ad8-75e0-4e2f-a3e7-373b3804a106": Phase="Pending", Reason="", readiness=false. Elapsed: 2.005758878s Jun 3 11:26:28.429: INFO: Pod "busybox-readonly-false-2cd70ad8-75e0-4e2f-a3e7-373b3804a106": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010509577s Jun 3 11:26:30.485: INFO: Pod "busybox-readonly-false-2cd70ad8-75e0-4e2f-a3e7-373b3804a106": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.066797505s Jun 3 11:26:30.485: INFO: Pod "busybox-readonly-false-2cd70ad8-75e0-4e2f-a3e7-373b3804a106" satisfied condition "Succeeded or Failed" [AfterEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:30.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-9798" for this suite. • [SLOW TEST:6.112 seconds] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 When creating a pod with readOnlyRootFilesystem /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:171 should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","total":-1,"completed":5,"skipped":117,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:28.073: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0666 on node default medium Jun 3 11:26:28.110: INFO: Waiting up to 5m0s for pod "pod-11a74dac-7a9c-46e9-b85b-ee73a355b592" in namespace "emptydir-3730" to be "Succeeded or Failed" Jun 3 11:26:28.112: INFO: Pod "pod-11a74dac-7a9c-46e9-b85b-ee73a355b592": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061028ms Jun 3 11:26:30.181: INFO: Pod "pod-11a74dac-7a9c-46e9-b85b-ee73a355b592": Phase="Pending", Reason="", readiness=false. Elapsed: 2.070928403s Jun 3 11:26:32.185: INFO: Pod "pod-11a74dac-7a9c-46e9-b85b-ee73a355b592": Phase="Running", Reason="", readiness=true. Elapsed: 4.074971215s Jun 3 11:26:34.282: INFO: Pod "pod-11a74dac-7a9c-46e9-b85b-ee73a355b592": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.171634876s STEP: Saw pod success Jun 3 11:26:34.282: INFO: Pod "pod-11a74dac-7a9c-46e9-b85b-ee73a355b592" satisfied condition "Succeeded or Failed" Jun 3 11:26:34.289: INFO: Trying to get logs from node v1.21-worker pod pod-11a74dac-7a9c-46e9-b85b-ee73a355b592 container test-container: STEP: delete the pod Jun 3 11:26:34.300: INFO: Waiting for pod pod-11a74dac-7a9c-46e9-b85b-ee73a355b592 to disappear Jun 3 11:26:34.302: INFO: Pod pod-11a74dac-7a9c-46e9-b85b-ee73a355b592 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:34.302: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3730" for this suite. • [SLOW TEST:6.236 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":155,"failed":0} [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:34.310: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name projected-configmap-test-volume-map-458e5576-ec55-4404-a234-d7918ed61b1f STEP: Creating a pod to test consume configMaps Jun 3 11:26:34.344: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a" in namespace "projected-6660" to be "Succeeded or Failed" Jun 3 11:26:34.347: INFO: Pod "pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.663512ms Jun 3 11:26:36.351: INFO: Pod "pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006715891s Jun 3 11:26:38.355: INFO: Pod "pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010778306s STEP: Saw pod success Jun 3 11:26:38.355: INFO: Pod "pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a" satisfied condition "Succeeded or Failed" Jun 3 11:26:38.358: INFO: Trying to get logs from node v1.21-worker pod pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a container agnhost-container: STEP: delete the pod Jun 3 11:26:38.373: INFO: Waiting for pod pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a to disappear Jun 3 11:26:38.379: INFO: Pod pod-projected-configmaps-56cd48d6-b05a-4d17-949b-1eb7bda27b8a no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:38.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6660" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":-1,"completed":5,"skipped":155,"failed":0} [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:38.391: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for the cluster [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1875.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1875.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:26:40.469: INFO: DNS probes using dns-1875/dns-test-5e93fb4e-e386-4bb1-880d-51092eef659f succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:40.479: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-1875" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for the cluster [Conformance]","total":-1,"completed":6,"skipped":155,"failed":0} SSS ------------------------------ [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:40.499: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:38 [It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:26:40.541: INFO: The status of Pod busybox-readonly-fsc766da4a-aebb-4c3a-9214-c4051e82f94c is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:42.546: INFO: The status of Pod busybox-readonly-fsc766da4a-aebb-4c3a-9214-c4051e82f94c is Running (Ready = true) [AfterEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:42.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-7317" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":158,"failed":0} SSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:16.888: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of same group and version but different kinds [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation Jun 3 11:26:17.489: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:26:24.159: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:44.492: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-6831" for this suite. • [SLOW TEST:27.614 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of same group and version but different kinds [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","total":-1,"completed":3,"skipped":55,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:42.574: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-volume-c9f8ad80-c212-435f-9667-4c6c2da919c4 STEP: Creating a pod to test consume configMaps Jun 3 11:26:42.622: INFO: Waiting up to 5m0s for pod "pod-configmaps-59a0225a-0519-460a-8fc0-d508ad7d1297" in namespace "configmap-132" to be "Succeeded or Failed" Jun 3 11:26:42.625: INFO: Pod "pod-configmaps-59a0225a-0519-460a-8fc0-d508ad7d1297": Phase="Pending", Reason="", readiness=false. Elapsed: 2.97319ms Jun 3 11:26:44.629: INFO: Pod "pod-configmaps-59a0225a-0519-460a-8fc0-d508ad7d1297": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.006827719s STEP: Saw pod success Jun 3 11:26:44.629: INFO: Pod "pod-configmaps-59a0225a-0519-460a-8fc0-d508ad7d1297" satisfied condition "Succeeded or Failed" Jun 3 11:26:44.632: INFO: Trying to get logs from node v1.21-worker pod pod-configmaps-59a0225a-0519-460a-8fc0-d508ad7d1297 container configmap-volume-test: STEP: delete the pod Jun 3 11:26:44.645: INFO: Waiting for pod pod-configmaps-59a0225a-0519-460a-8fc0-d508ad7d1297 to disappear Jun 3 11:26:44.648: INFO: Pod pod-configmaps-59a0225a-0519-460a-8fc0-d508ad7d1297 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:44.648: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-132" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":161,"failed":0} SSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:23.925: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should have session affinity work for NodePort service [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service in namespace services-3236 STEP: creating service affinity-nodeport in namespace services-3236 STEP: creating replication controller affinity-nodeport in namespace services-3236 I0603 11:26:23.971833 21 runners.go:190] Created replication controller with name: affinity-nodeport, namespace: services-3236, replica count: 3 I0603 11:26:27.022732 21 runners.go:190] affinity-nodeport Pods: 3 out of 3 created, 1 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:26:30.023531 21 runners.go:190] affinity-nodeport Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:26:30.034: INFO: Creating new exec pod Jun 3 11:26:35.052: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3236 exec execpod-affinitymkb8z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport 80' Jun 3 11:26:35.310: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport 80\nConnection to affinity-nodeport 80 port [tcp/http] succeeded!\n" Jun 3 11:26:35.311: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:26:35.311: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3236 exec execpod-affinitymkb8z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.139.74 80' Jun 3 11:26:35.574: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.139.74 80\nConnection to 10.96.139.74 80 port [tcp/http] succeeded!\n" Jun 3 11:26:35.574: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:26:35.575: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3236 exec execpod-affinitymkb8z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.4 31575' Jun 3 11:26:35.780: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.4 31575\nConnection to 172.18.0.4 31575 port [tcp/*] succeeded!\n" Jun 3 11:26:35.780: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:26:35.780: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3236 exec execpod-affinitymkb8z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.2 31575' Jun 3 11:26:36.110: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.2 31575\nConnection to 172.18.0.2 31575 port [tcp/*] succeeded!\n" Jun 3 11:26:36.110: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:26:36.111: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3236 exec execpod-affinitymkb8z -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.4:31575/ ; done' Jun 3 11:26:36.457: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:31575/\n" Jun 3 11:26:36.457: INFO: stdout: "\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw\naffinity-nodeport-vlxcw" Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Received response from host: affinity-nodeport-vlxcw Jun 3 11:26:36.457: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-nodeport in namespace services-3236, will wait for the garbage collector to delete the pods Jun 3 11:26:36.541: INFO: Deleting ReplicationController affinity-nodeport took: 4.471129ms Jun 3 11:26:36.641: INFO: Terminating ReplicationController affinity-nodeport pods took: 100.830766ms [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:47.488: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3236" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:23.577 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should have session affinity work for NodePort service [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance]","total":-1,"completed":4,"skipped":47,"failed":0} SS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:47.507: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should check if Kubernetes control plane services is included in cluster-info [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: validating cluster-info Jun 3 11:26:47.537: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5084 cluster-info' Jun 3 11:26:47.659: INFO: stderr: "" Jun 3 11:26:47.659: INFO: stdout: "\x1b[0;32mKubernetes control plane\x1b[0m is running at \x1b[0;33mhttps://172.30.13.90:33295\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:47.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5084" for this suite. • ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes control plane services is included in cluster-info [Conformance]","total":-1,"completed":5,"skipped":49,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:20.210: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105 STEP: Creating service test in namespace statefulset-5711 [It] Should recreate evicted statefulset [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Looking for a node to schedule stateful set and pod STEP: Creating pod with conflicting port in namespace statefulset-5711 STEP: Creating statefulset with conflicting port in namespace statefulset-5711 STEP: Waiting until pod test-pod will start running in namespace statefulset-5711 STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-5711 Jun 3 11:26:24.271: INFO: Observed stateful pod in namespace: statefulset-5711, name: ss-0, uid: 5ba1a773-5c7a-4af2-8cd8-327471620e91, status phase: Failed. Waiting for statefulset controller to delete. Jun 3 11:26:24.274: INFO: Observed stateful pod in namespace: statefulset-5711, name: ss-0, uid: 5ba1a773-5c7a-4af2-8cd8-327471620e91, status phase: Failed. Waiting for statefulset controller to delete. Jun 3 11:26:24.275: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-5711 STEP: Removing pod with conflicting port in namespace statefulset-5711 STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-5711 and will be in running state [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116 Jun 3 11:26:32.386: INFO: Deleting all statefulset in ns statefulset-5711 Jun 3 11:26:32.389: INFO: Scaling statefulset ss to 0 Jun 3 11:26:52.405: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:26:52.409: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:52.424: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-5711" for this suite. • [SLOW TEST:32.222 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95 Should recreate evicted statefulset [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","total":-1,"completed":5,"skipped":70,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:44.517: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a job STEP: Ensuring job reaches completions [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:54.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-9062" for this suite. • [SLOW TEST:10.046 seconds] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":-1,"completed":4,"skipped":61,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:54.579: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context STEP: Waiting for a default service account to be provisioned in namespace [It] should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser Jun 3 11:26:54.619: INFO: Waiting up to 5m0s for pod "security-context-d6c30cec-1b97-415b-b052-e2032cfd8124" in namespace "security-context-8852" to be "Succeeded or Failed" Jun 3 11:26:54.622: INFO: Pod "security-context-d6c30cec-1b97-415b-b052-e2032cfd8124": Phase="Pending", Reason="", readiness=false. Elapsed: 2.622677ms Jun 3 11:26:56.626: INFO: Pod "security-context-d6c30cec-1b97-415b-b052-e2032cfd8124": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007007782s Jun 3 11:26:58.631: INFO: Pod "security-context-d6c30cec-1b97-415b-b052-e2032cfd8124": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011712025s STEP: Saw pod success Jun 3 11:26:58.631: INFO: Pod "security-context-d6c30cec-1b97-415b-b052-e2032cfd8124" satisfied condition "Succeeded or Failed" Jun 3 11:26:58.634: INFO: Trying to get logs from node v1.21-worker pod security-context-d6c30cec-1b97-415b-b052-e2032cfd8124 container test-container: STEP: delete the pod Jun 3 11:26:58.650: INFO: Waiting for pod security-context-d6c30cec-1b97-415b-b052-e2032cfd8124 to disappear Jun 3 11:26:58.654: INFO: Pod security-context-d6c30cec-1b97-415b-b052-e2032cfd8124 no longer exists [AfterEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:58.654: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-8852" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]","total":-1,"completed":5,"skipped":67,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:28.737: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] updates the published spec when one version gets renamed [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: set up a multi version CRD Jun 3 11:26:28.770: INFO: >>> kubeConfig: /root/.kube/config STEP: rename a version STEP: check the new version name is served STEP: check the old version name is removed STEP: check the other version is not changed [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:58.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-9390" for this suite. • [SLOW TEST:29.939 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 updates the published spec when one version gets renamed [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","total":-1,"completed":5,"skipped":188,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Aggregator /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:52.518: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename aggregator STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Aggregator /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:77 Jun 3 11:26:52.552: INFO: >>> kubeConfig: /root/.kube/config [It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering the sample API server. Jun 3 11:26:53.438: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set Jun 3 11:26:55.471: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316413, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316413, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316413, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316413, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-64f6b9dc99\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:26:58.504: INFO: Waited 1.023840371s for the sample-apiserver to be ready to handle requests. STEP: Read Status for v1alpha1.wardle.example.com STEP: kubectl patch apiservice v1alpha1.wardle.example.com -p '{"spec":{"versionPriority": 400}}' STEP: List APIServices Jun 3 11:26:58.893: INFO: Found v1alpha1.wardle.example.com in APIServiceList [AfterEach] [sig-api-machinery] Aggregator /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:68 [AfterEach] [sig-api-machinery] Aggregator /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:26:59.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "aggregator-5583" for this suite. • [SLOW TEST:7.221 seconds] [sig-api-machinery] Aggregator /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]","total":-1,"completed":6,"skipped":114,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:57.927: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete RS created by deployment when not orphaning [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for all rs to be garbage collected STEP: expected 0 rs, got 1 rs STEP: expected 0 pods, got 2 pods STEP: Gathering metrics W0603 11:25:59.029629 34 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Jun 3 11:27:01.048: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:01.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-8653" for this suite. • [SLOW TEST:63.131 seconds] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete RS created by deployment when not orphaning [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":-1,"completed":3,"skipped":68,"failed":0} SS ------------------------------ [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:01.063: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename endpointslice STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/endpointslice.go:49 [It] should have Endpoints and EndpointSlices pointing to API Server [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:01.100: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "endpointslice-345" for this suite. • ------------------------------ [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:01.888: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:01.943: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-9018" for this suite. • [SLOW TEST:60.070 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":28,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:47.695: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should verify ResourceQuota with terminating scopes. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a ResourceQuota with terminating scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a ResourceQuota with not terminating scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a long running pod STEP: Ensuring resource quota with not terminating scope captures the pod usage STEP: Ensuring resource quota with terminating scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage STEP: Creating a terminating pod STEP: Ensuring resource quota with terminating scope captures the pod usage STEP: Ensuring resource quota with not terminating scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:03.915: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-9827" for this suite. • [SLOW TEST:16.228 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should verify ResourceQuota with terminating scopes. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","total":-1,"completed":6,"skipped":62,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:58.723: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-e64a2743-0c9c-4da9-a4f7-aa61408f7235 STEP: Creating a pod to test consume secrets Jun 3 11:26:58.882: INFO: Waiting up to 5m0s for pod "pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280" in namespace "secrets-2279" to be "Succeeded or Failed" Jun 3 11:26:58.886: INFO: Pod "pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280": Phase="Pending", Reason="", readiness=false. Elapsed: 3.910932ms Jun 3 11:27:00.890: INFO: Pod "pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008573868s Jun 3 11:27:02.895: INFO: Pod "pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013017108s Jun 3 11:27:04.898: INFO: Pod "pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015989608s Jun 3 11:27:06.902: INFO: Pod "pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.020451116s STEP: Saw pod success Jun 3 11:27:06.902: INFO: Pod "pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280" satisfied condition "Succeeded or Failed" Jun 3 11:27:06.905: INFO: Trying to get logs from node v1.21-worker pod pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280 container secret-volume-test: STEP: delete the pod Jun 3 11:27:06.920: INFO: Waiting for pod pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280 to disappear Jun 3 11:27:06.923: INFO: Pod pod-secrets-fe24e2a5-a960-4a34-a345-eeb888b82280 no longer exists [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:06.923: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-2279" for this suite. STEP: Destroying namespace "secret-namespace-2057" for this suite. • [SLOW TEST:8.215 seconds] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":-1,"completed":6,"skipped":216,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:03.976: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0666 on tmpfs Jun 3 11:27:04.010: INFO: Waiting up to 5m0s for pod "pod-c24f844a-5e2a-473b-903e-881a88140492" in namespace "emptydir-5974" to be "Succeeded or Failed" Jun 3 11:27:04.013: INFO: Pod "pod-c24f844a-5e2a-473b-903e-881a88140492": Phase="Pending", Reason="", readiness=false. Elapsed: 2.714085ms Jun 3 11:27:06.018: INFO: Pod "pod-c24f844a-5e2a-473b-903e-881a88140492": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007574814s Jun 3 11:27:08.022: INFO: Pod "pod-c24f844a-5e2a-473b-903e-881a88140492": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011516872s Jun 3 11:27:10.027: INFO: Pod "pod-c24f844a-5e2a-473b-903e-881a88140492": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016415395s STEP: Saw pod success Jun 3 11:27:10.027: INFO: Pod "pod-c24f844a-5e2a-473b-903e-881a88140492" satisfied condition "Succeeded or Failed" Jun 3 11:27:10.030: INFO: Trying to get logs from node v1.21-worker2 pod pod-c24f844a-5e2a-473b-903e-881a88140492 container test-container: STEP: delete the pod Jun 3 11:27:10.425: INFO: Waiting for pod pod-c24f844a-5e2a-473b-903e-881a88140492 to disappear Jun 3 11:27:10.428: INFO: Pod pod-c24f844a-5e2a-473b-903e-881a88140492 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:10.428: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-5974" for this suite. • [SLOW TEST:6.462 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":93,"failed":0} S ------------------------------ {"msg":"PASSED [sig-network] EndpointSlice should have Endpoints and EndpointSlices pointing to API Server [Conformance]","total":-1,"completed":4,"skipped":70,"failed":0} [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:01.110: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:86 [It] deployment should support proportional scaling [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:27:01.138: INFO: Creating deployment "webserver-deployment" Jun 3 11:27:01.143: INFO: Waiting for observed generation 1 Jun 3 11:27:03.149: INFO: Waiting for all required pods to come up Jun 3 11:27:03.153: INFO: Pod name httpd: Found 10 pods out of 10 STEP: ensuring each pod is running Jun 3 11:27:13.161: INFO: Waiting for deployment "webserver-deployment" to complete Jun 3 11:27:13.168: INFO: Updating deployment "webserver-deployment" with a non-existent image Jun 3 11:27:13.178: INFO: Updating deployment webserver-deployment Jun 3 11:27:13.178: INFO: Waiting for observed generation 2 Jun 3 11:27:15.186: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 Jun 3 11:27:15.189: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 Jun 3 11:27:15.193: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas Jun 3 11:27:15.202: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 Jun 3 11:27:15.202: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 Jun 3 11:27:15.205: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas Jun 3 11:27:15.211: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas Jun 3 11:27:15.211: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 Jun 3 11:27:15.220: INFO: Updating deployment webserver-deployment Jun 3 11:27:15.220: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas Jun 3 11:27:15.226: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 Jun 3 11:27:15.229: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:80 Jun 3 11:27:17.238: INFO: Deployment "webserver-deployment": &Deployment{ObjectMeta:{webserver-deployment deployment-521 c01b06c3-a763-4abd-af32-90623e70d9db 3286772 3 2021-06-03 11:27:01 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004d2b1b8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2021-06-03 11:27:15 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-795d758f88" is progressing.,LastUpdateTime:2021-06-03 11:27:15 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} Jun 3 11:27:17.242: INFO: New ReplicaSet "webserver-deployment-795d758f88" of Deployment "webserver-deployment": &ReplicaSet{ObjectMeta:{webserver-deployment-795d758f88 deployment-521 3362a292-efa8-4df6-9342-ad7f7afaa547 3286770 3 2021-06-03 11:27:13 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment c01b06c3-a763-4abd-af32-90623e70d9db 0xc004d2b587 0xc004d2b588}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c01b06c3-a763-4abd-af32-90623e70d9db\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 795d758f88,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004d2b608 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:27:17.242: INFO: All old ReplicaSets of Deployment "webserver-deployment": Jun 3 11:27:17.242: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-847dcfb7fb deployment-521 f45303c4-4148-43d4-aab0-658e1a99abee 3286760 3 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment c01b06c3-a763-4abd-af32-90623e70d9db 0xc004d2b667 0xc004d2b668}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:27:06 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c01b06c3-a763-4abd-af32-90623e70d9db\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 847dcfb7fb,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004d2b6d8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:27:17.251: INFO: Pod "webserver-deployment-795d758f88-5mklb" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-5mklb webserver-deployment-795d758f88- deployment-521 7fe9f63d-b8b5-48fd-a223-42be980c4c38 3286752 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc004d2bb47 0xc004d2bb48}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-l6cpb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6cpb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.251: INFO: Pod "webserver-deployment-795d758f88-5rlh9" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-5rlh9 webserver-deployment-795d758f88- deployment-521 318f3096-9a53-4220-8dc8-485d3d7f7c32 3286748 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc004d2bcc7 0xc004d2bcc8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-swd4b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-swd4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.252: INFO: Pod "webserver-deployment-795d758f88-684ls" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-684ls webserver-deployment-795d758f88- deployment-521 daabd533-9654-40e7-9d80-423f2c0a0223 3286830 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.204" ], "mac": "62:1e:23:72:8a:bf", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.204" ], "mac": "62:1e:23:72:8a:bf", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc004d2be47 0xc004d2be48}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wff7z,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wff7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.252: INFO: Pod "webserver-deployment-795d758f88-cbm2w" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-cbm2w webserver-deployment-795d758f88- deployment-521 560d6530-6a55-4238-97b5-1578f927622f 3286849 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.244" ], "mac": "92:f9:2c:8a:83:cf", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.244" ], "mac": "92:f9:2c:8a:83:cf", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc004d2bfd7 0xc004d2bfd8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fflh5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fflh5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.252: INFO: Pod "webserver-deployment-795d758f88-gh2nd" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-gh2nd webserver-deployment-795d758f88- deployment-521 2bebf19f-29a4-4911-973b-f9f64a1fc512 3286674 0 2021-06-03 11:27:13 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.198" ], "mac": "3a:ce:9f:bb:99:3e", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.198" ], "mac": "3a:ce:9f:bb:99:3e", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d12167 0xc003d12168}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:14 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-kr6m4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kr6m4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.252: INFO: Pod "webserver-deployment-795d758f88-hg24s" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-hg24s webserver-deployment-795d758f88- deployment-521 28900bf1-cf23-4f47-b886-ec1c7d94a5e0 3286773 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d122f7 0xc003d122f8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-j5gdd,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j5gdd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:15 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.253: INFO: Pod "webserver-deployment-795d758f88-jtgn4" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-jtgn4 webserver-deployment-795d758f88- deployment-521 17591812-fc30-420b-905b-966013631feb 3286676 0 2021-06-03 11:27:13 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.199" ], "mac": "92:f5:94:9a:f4:67", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.199" ], "mac": "92:f5:94:9a:f4:67", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d124d7 0xc003d124d8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:14 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-rknql,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rknql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.253: INFO: Pod "webserver-deployment-795d758f88-kk6w8" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-kk6w8 webserver-deployment-795d758f88- deployment-521 50ba7956-5e68-4dde-93fc-d5cf9582b5ad 3286663 0 2021-06-03 11:27:13 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.238" ], "mac": "ce:6d:dd:21:69:43", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.238" ], "mac": "ce:6d:dd:21:69:43", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d12667 0xc003d12668}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}} {multus Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-4hl9n,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4hl9n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:13 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.254: INFO: Pod "webserver-deployment-795d758f88-kx6w5" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-kx6w5 webserver-deployment-795d758f88- deployment-521 dbd24360-d685-48dd-95b3-9dcf3fbef4cc 3286831 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.242" ], "mac": "ea:b5:c4:eb:a5:3a", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.242" ], "mac": "ea:b5:c4:eb:a5:3a", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d12867 0xc003d12868}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-67xrn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-67xrn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:15 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.254: INFO: Pod "webserver-deployment-795d758f88-p9fq8" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-p9fq8 webserver-deployment-795d758f88- deployment-521 225ae2f0-f733-41b4-9b10-817c94f6431f 3286679 0 2021-06-03 11:27:13 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.200" ], "mac": "c6:9f:fe:79:ff:f6", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.200" ], "mac": "c6:9f:fe:79:ff:f6", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d12a67 0xc003d12a68}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:14 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fbxch,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fbxch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.254: INFO: Pod "webserver-deployment-795d758f88-x9lfh" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-x9lfh webserver-deployment-795d758f88- deployment-521 48ca2f8c-9078-406d-91ea-cf774d6083b0 3286821 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.202" ], "mac": "4a:22:bf:b9:77:56", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.202" ], "mac": "4a:22:bf:b9:77:56", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d12bf7 0xc003d12bf8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-jv72k,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jv72k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.254: INFO: Pod "webserver-deployment-795d758f88-xrrpl" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-xrrpl webserver-deployment-795d758f88- deployment-521 9dd94d61-9db7-4f3e-82fc-b123ef9a00c5 3286665 0 2021-06-03 11:27:13 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.239" ], "mac": "ca:94:c8:2b:6b:bb", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.239" ], "mac": "ca:94:c8:2b:6b:bb", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d12d87 0xc003d12d88}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}} {multus Update v1 2021-06-03 11:27:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-ljw8m,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ljw8m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:13 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.255: INFO: Pod "webserver-deployment-795d758f88-zch84" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-zch84 webserver-deployment-795d758f88- deployment-521 7bc506bb-2ae5-4a46-a982-673d4d434033 3286854 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.245" ], "mac": "92:e6:b0:2d:17:7e", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.245" ], "mac": "92:e6:b0:2d:17:7e", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 3362a292-efa8-4df6-9342-ad7f7afaa547 0xc003d12f87 0xc003d12f88}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"3362a292-efa8-4df6-9342-ad7f7afaa547\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}} {multus Update v1 2021-06-03 11:27:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-k7j4b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k7j4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:15 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.255: INFO: Pod "webserver-deployment-847dcfb7fb-2dgp7" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-2dgp7 webserver-deployment-847dcfb7fb- deployment-521 771deae0-520b-4740-96d7-d5044fcf84e7 3286736 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d13187 0xc003d13188}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-gcrfj,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gcrfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.255: INFO: Pod "webserver-deployment-847dcfb7fb-4sbsf" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-4sbsf webserver-deployment-847dcfb7fb- deployment-521 336dbd48-a324-4342-b958-ed83d878b7db 3286811 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.201" ], "mac": "ee:cb:e2:4b:7e:b4", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.201" ], "mac": "ee:cb:e2:4b:7e:b4", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d132f7 0xc003d132f8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-8xxjk,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8xxjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.256: INFO: Pod "webserver-deployment-847dcfb7fb-69khn" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-69khn webserver-deployment-847dcfb7fb- deployment-521 175e19ec-044c-4e07-98a1-d6f56b7be9c0 3286734 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d13477 0xc003d13478}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-7d9d8,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7d9d8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.256: INFO: Pod "webserver-deployment-847dcfb7fb-6j6ct" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-6j6ct webserver-deployment-847dcfb7fb- deployment-521 9b76c2d1-e5d2-43b2-9581-baabe25c3dc0 3286818 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.240" ], "mac": "b2:79:2a:a1:1e:65", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.240" ], "mac": "b2:79:2a:a1:1e:65", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d135e7 0xc003d135e8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-kqrlk,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kqrlk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:15 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.256: INFO: Pod "webserver-deployment-847dcfb7fb-7rgc6" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-7rgc6 webserver-deployment-847dcfb7fb- deployment-521 875fe619-860d-4705-a879-242755e288f6 3286487 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.188" ], "mac": "f2:f4:89:25:8f:f5", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.188" ], "mac": "f2:f4:89:25:8f:f5", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d137c7 0xc003d137c8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.188\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-7nhst,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7nhst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.4,PodIP:10.244.1.188,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://4848bad2d721fff1703ed934504916036f25d69444d73e27232456b108fd73c5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.188,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.257: INFO: Pod "webserver-deployment-847dcfb7fb-85bms" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-85bms webserver-deployment-847dcfb7fb- deployment-521 217a7ffa-65e6-4944-a8a4-653eaa2520e4 3286718 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d139c7 0xc003d139c8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-5ft8m,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5ft8m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.257: INFO: Pod "webserver-deployment-847dcfb7fb-dfsbk" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-dfsbk webserver-deployment-847dcfb7fb- deployment-521 4090631b-c0c5-4097-95d2-22128c63efa4 3286398 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.231" ], "mac": "6a:30:da:c4:f8:4d", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.231" ], "mac": "6a:30:da:c4:f8:4d", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d13b37 0xc003d13b38}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.231\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-tmpgp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tmpgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:10.244.2.231,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://463742060e345a401bad5421272112ba604b3ef066cf7ea77439e0305d2f8e7e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.231,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.257: INFO: Pod "webserver-deployment-847dcfb7fb-dh58g" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-dh58g webserver-deployment-847dcfb7fb- deployment-521 c88082ca-6cb9-4823-bbf1-6fe90fbac207 3286836 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.205" ], "mac": "96:71:02:d3:f5:44", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.205" ], "mac": "96:71:02:d3:f5:44", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d13d37 0xc003d13d38}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-97pwv,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-97pwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.259: INFO: Pod "webserver-deployment-847dcfb7fb-fqnmm" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-fqnmm webserver-deployment-847dcfb7fb- deployment-521 76892060-d4a5-496e-8378-fce4b27245a9 3286439 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.234" ], "mac": "4a:88:33:1f:40:5b", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.234" ], "mac": "4a:88:33:1f:40:5b", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d13eb7 0xc003d13eb8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.234\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-25rnr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-25rnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:10.244.2.234,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://13519c65758fb4f3b7646cde47594e73be320b4f745756d687c7b699ba85b35b,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.234,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.259: INFO: Pod "webserver-deployment-847dcfb7fb-fts8b" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-fts8b webserver-deployment-847dcfb7fb- deployment-521 39d63db7-2988-4dca-8517-8800aaa858f5 3286832 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.241" ], "mac": "12:84:04:05:a4:9a", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.241" ], "mac": "12:84:04:05:a4:9a", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d620b7 0xc003d620b8}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-5pzmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5pzmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:15 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.259: INFO: Pod "webserver-deployment-847dcfb7fb-h49gg" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-h49gg webserver-deployment-847dcfb7fb- deployment-521 d558db96-4ace-48ed-acb2-44782a661aa8 3286494 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.232" ], "mac": "aa:a0:26:7b:07:81", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.232" ], "mac": "aa:a0:26:7b:07:81", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62297 0xc003d62298}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.232\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-295g5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-295g5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:10.244.2.232,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://e39fd03879b47a0a1ea54ec9a31955e668dd31fcf989fdfabb6c85730cad8953,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.232,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.259: INFO: Pod "webserver-deployment-847dcfb7fb-kcgq8" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-kcgq8 webserver-deployment-847dcfb7fb- deployment-521 f46be8ad-f8d6-4db9-b96e-816658e7ab09 3286856 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.207" ], "mac": "ce:a2:06:2e:7c:74", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.207" ], "mac": "ce:a2:06:2e:7c:74", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62497 0xc003d62498}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-d98nt,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d98nt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.260: INFO: Pod "webserver-deployment-847dcfb7fb-kt6w8" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-kt6w8 webserver-deployment-847dcfb7fb- deployment-521 e79e0d26-f355-4eae-abde-3d09f0743b65 3286850 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.206" ], "mac": "96:ff:7d:10:90:53", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.206" ], "mac": "96:ff:7d:10:90:53", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62617 0xc003d62618}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-hj6bx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hj6bx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.260: INFO: Pod "webserver-deployment-847dcfb7fb-kxwj5" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-kxwj5 webserver-deployment-847dcfb7fb- deployment-521 dd2d9fa2-efd3-4e05-b328-17d79e7a358c 3286834 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.203" ], "mac": "d6:c3:76:71:a4:3b", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.203" ], "mac": "d6:c3:76:71:a4:3b", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62797 0xc003d62798}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fnjkk,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fnjkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.260: INFO: Pod "webserver-deployment-847dcfb7fb-lk6pq" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-lk6pq webserver-deployment-847dcfb7fb- deployment-521 9ff317cd-1c28-4bc3-a469-e04ec3c6c365 3286727 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62917 0xc003d62918}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-5hq8g,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5hq8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.260: INFO: Pod "webserver-deployment-847dcfb7fb-lrd9w" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-lrd9w webserver-deployment-847dcfb7fb- deployment-521 b2134c8d-3f36-4e11-9e28-45ac15729b52 3286533 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.189" ], "mac": "fe:fc:4e:3d:07:d9", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.189" ], "mac": "fe:fc:4e:3d:07:d9", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62a87 0xc003d62a88}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.189\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-m2xqs,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2xqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.4,PodIP:10.244.1.189,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://ac13872d9393029e1b4b38086261f15c2f13e0ba3adffe12b598979d425df2f6,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.189,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.261: INFO: Pod "webserver-deployment-847dcfb7fb-nd7x6" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-nd7x6 webserver-deployment-847dcfb7fb- deployment-521 7a912044-180a-4854-b08c-db1436343628 3286481 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.235" ], "mac": "42:d9:d8:e8:cc:fa", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.235" ], "mac": "42:d9:d8:e8:cc:fa", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62c87 0xc003d62c88}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.235\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-smh8v,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-smh8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:10.244.2.235,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://52cd37c202e75e73bc5f6c5cd0d6bd45afccd3539cd52f6e4934418c14483438,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.235,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.261: INFO: Pod "webserver-deployment-847dcfb7fb-p5r8b" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-p5r8b webserver-deployment-847dcfb7fb- deployment-521 d50cd06e-afe9-4699-b331-872da5c0c156 3286413 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.233" ], "mac": "da:84:3d:93:5b:25", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.233" ], "mac": "da:84:3d:93:5b:25", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d62e87 0xc003d62e88}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.233\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-qx4f9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qx4f9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:10.244.2.233,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://e71f0e0df5b4d80eb46948dac70f5c65ff769581eb2eeebc6f87efee12c7a050,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.233,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.261: INFO: Pod "webserver-deployment-847dcfb7fb-tnmgg" is not available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-tnmgg webserver-deployment-847dcfb7fb- deployment-521 33654273-0604-4145-82e6-d4904a5157aa 3286842 0 2021-06-03 11:27:15 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.243" ], "mac": "42:e6:8d:d7:eb:6d", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.243" ], "mac": "42:e6:8d:d7:eb:6d", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d63087 0xc003d63088}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2021-06-03 11:27:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}} {multus Update v1 2021-06-03 11:27:16 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-2j4ld,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2j4ld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:,StartTime:2021-06-03 11:27:15 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:27:17.261: INFO: Pod "webserver-deployment-847dcfb7fb-wdbzz" is available: &Pod{ObjectMeta:{webserver-deployment-847dcfb7fb-wdbzz webserver-deployment-847dcfb7fb- deployment-521 c30a7e43-06d7-4ef3-a21e-970d434c1962 3286549 0 2021-06-03 11:27:01 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.192" ], "mac": "fa:bd:a9:a9:2f:42", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.192" ], "mac": "fa:bd:a9:a9:2f:42", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-deployment-847dcfb7fb f45303c4-4148-43d4-aab0-658e1a99abee 0xc003d63267 0xc003d63268}] [] [{kube-controller-manager Update v1 2021-06-03 11:27:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f45303c4-4148-43d4-aab0-658e1a99abee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:27:03 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:27:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.192\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-rtmgc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rtmgc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:27:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.4,PodIP:10.244.1.192,StartTime:2021-06-03 11:27:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:27:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://855e7e8bccd954e2106d8120aeb254e2efdfcb8def19f29affe97ac986a9593f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.192,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:17.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-521" for this suite. • [SLOW TEST:16.159 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support proportional scaling [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":-1,"completed":5,"skipped":70,"failed":0} SS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:58.758: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should be able to change the type from ExternalName to ClusterIP [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a service externalname-service with the type=ExternalName in namespace services-19 STEP: changing the ExternalName service to type=ClusterIP STEP: creating replication controller externalname-service in namespace services-19 I0603 11:26:58.895598 26 runners.go:190] Created replication controller with name: externalname-service, namespace: services-19, replica count: 2 I0603 11:27:01.946952 26 runners.go:190] externalname-service Pods: 2 out of 2 created, 1 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:27:04.947167 26 runners.go:190] externalname-service Pods: 2 out of 2 created, 1 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:27:07.948291 26 runners.go:190] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:27:07.948: INFO: Creating new exec pod Jun 3 11:27:23.001: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-19 exec execpodp5c86 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' Jun 3 11:27:23.203: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" Jun 3 11:27:23.203: INFO: stdout: "externalname-service-xh74d" Jun 3 11:27:23.203: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-19 exec execpodp5c86 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.7.121 80' Jun 3 11:27:23.449: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.7.121 80\nConnection to 10.96.7.121 80 port [tcp/http] succeeded!\n" Jun 3 11:27:23.449: INFO: stdout: "externalname-service-h9phg" Jun 3 11:27:23.449: INFO: Cleaning up the ExternalName to ClusterIP test service [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:23.460: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-19" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:24.712 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should be able to change the type from ExternalName to ClusterIP [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","total":-1,"completed":6,"skipped":120,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir wrapper volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:17.285: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not conflict [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:27:17.343: INFO: The status of Pod pod-secrets-75bd6e32-5ac5-4033-af20-5b16746c6194 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:19.347: INFO: The status of Pod pod-secrets-75bd6e32-5ac5-4033-af20-5b16746c6194 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:21.348: INFO: The status of Pod pod-secrets-75bd6e32-5ac5-4033-af20-5b16746c6194 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:23.348: INFO: The status of Pod pod-secrets-75bd6e32-5ac5-4033-af20-5b16746c6194 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:25.348: INFO: The status of Pod pod-secrets-75bd6e32-5ac5-4033-af20-5b16746c6194 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:27.348: INFO: The status of Pod pod-secrets-75bd6e32-5ac5-4033-af20-5b16746c6194 is Running (Ready = true) STEP: Cleaning up the secret STEP: Cleaning up the configmap STEP: Cleaning up the pod [AfterEach] [sig-storage] EmptyDir wrapper volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:27.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-9316" for this suite. • [SLOW TEST:10.092 seconds] [sig-storage] EmptyDir wrapper volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 should not conflict [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":-1,"completed":6,"skipped":72,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:27.487: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create secret due to empty secret key [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating projection with secret that has name secret-emptykey-test-8b07679b-63fd-491c-bf34-52d0c19a189c [AfterEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:27.523: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-5403" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Secrets should fail to create secret due to empty secret key [Conformance]","total":-1,"completed":7,"skipped":139,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:24.438: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod busybox-dd571fc3-3662-4976-b4e4-53a172b4e91d in namespace container-probe-7294 Jun 3 11:26:30.500: INFO: Started pod busybox-dd571fc3-3662-4976-b4e4-53a172b4e91d in namespace container-probe-7294 STEP: checking the pod's current state and verifying that restartCount is present Jun 3 11:26:30.503: INFO: Initial restart count of pod busybox-dd571fc3-3662-4976-b4e4-53a172b4e91d is 0 Jun 3 11:27:28.678: INFO: Restart count of pod container-probe-7294/busybox-dd571fc3-3662-4976-b4e4-53a172b4e91d is now 1 (58.174902745s elapsed) STEP: deleting the pod [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:28.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-7294" for this suite. • [SLOW TEST:64.356 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":-1,"completed":3,"skipped":91,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:44.677: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: udp [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Performing setup for networking test in namespace pod-network-test-6328 STEP: creating a selector STEP: Creating the service pods in kubernetes Jun 3 11:26:44.706: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Jun 3 11:26:44.725: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:46.729: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:26:48.731: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:26:50.729: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:26:52.729: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:26:54.730: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:26:56.729: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:26:58.729: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:27:00.730: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:27:02.731: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:27:04.730: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:27:06.730: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:27:08.731: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:27:10.730: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:27:12.730: INFO: The status of Pod netserver-0 is Running (Ready = true) Jun 3 11:27:12.737: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Jun 3 11:27:30.756: INFO: Setting MaxTries for pod polling to 34 for networking test based on endpoint count 2 Jun 3 11:27:30.756: INFO: Breadth first check of 10.244.1.174 on host 172.18.0.4... Jun 3 11:27:30.760: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.197:9080/dial?request=hostname&protocol=udp&host=10.244.1.174&port=8081&tries=1'] Namespace:pod-network-test-6328 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:30.760: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:30.890: INFO: Waiting for responses: map[] Jun 3 11:27:30.890: INFO: reached 10.244.1.174 after 0/1 tries Jun 3 11:27:30.890: INFO: Breadth first check of 10.244.2.224 on host 172.18.0.2... Jun 3 11:27:30.893: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.197:9080/dial?request=hostname&protocol=udp&host=10.244.2.224&port=8081&tries=1'] Namespace:pod-network-test-6328 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:30.893: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:31.007: INFO: Waiting for responses: map[] Jun 3 11:27:31.008: INFO: reached 10.244.2.224 after 0/1 tries Jun 3 11:27:31.008: INFO: Going to retry 0 out of 2 pods.... [AfterEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:31.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-6328" for this suite. • [SLOW TEST:46.340 seconds] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/framework.go:23 Granular Checks: Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/networking.go:30 should function for intra-pod communication: udp [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":172,"failed":0} S ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:31.022: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be immutable if `immutable` field is set [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:31.089: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-5588" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be immutable if `immutable` field is set [Conformance]","total":-1,"completed":10,"skipped":173,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:27.635: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context STEP: Waiting for a default service account to be provisioned in namespace [It] should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser Jun 3 11:27:27.672: INFO: Waiting up to 5m0s for pod "security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0" in namespace "security-context-6503" to be "Succeeded or Failed" Jun 3 11:27:27.675: INFO: Pod "security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.821018ms Jun 3 11:27:29.679: INFO: Pod "security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006949767s Jun 3 11:27:31.684: INFO: Pod "security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011083937s Jun 3 11:27:33.687: INFO: Pod "security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.014351961s STEP: Saw pod success Jun 3 11:27:33.687: INFO: Pod "security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0" satisfied condition "Succeeded or Failed" Jun 3 11:27:33.690: INFO: Trying to get logs from node v1.21-worker pod security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0 container test-container: STEP: delete the pod Jun 3 11:27:33.710: INFO: Waiting for pod security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0 to disappear Jun 3 11:27:33.713: INFO: Pod security-context-3428ca92-338f-4ac8-83f7-13d25ff438e0 no longer exists [AfterEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:33.713: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-6503" for this suite. • [SLOW TEST:6.086 seconds] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23 should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Security Context should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]","total":-1,"completed":8,"skipped":191,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:31.156: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should mount an API token into pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: getting the auto-created API token STEP: reading a file in the container Jun 3 11:27:37.717: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4686 pod-service-account-c6c31d2b-0dcf-47d9-93a3-0d3bc85dbdaa -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' STEP: reading a file in the container Jun 3 11:27:37.975: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4686 pod-service-account-c6c31d2b-0dcf-47d9-93a3-0d3bc85dbdaa -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' STEP: reading a file in the container Jun 3 11:27:38.267: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4686 pod-service-account-c6c31d2b-0dcf-47d9-93a3-0d3bc85dbdaa -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' [AfterEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:38.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-4686" for this suite. • [SLOW TEST:7.395 seconds] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should mount an API token into pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods [Conformance]","total":-1,"completed":11,"skipped":207,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:33.799: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/init_container.go:162 [It] should invoke init containers on a RestartAlways pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod Jun 3 11:27:33.827: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:40.819: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-9872" for this suite. • [SLOW TEST:7.380 seconds] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should invoke init containers on a RestartAlways pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","total":-1,"completed":9,"skipped":229,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] EndpointSliceMirroring /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:38.568: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename endpointslicemirroring STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] EndpointSliceMirroring /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/endpointslicemirroring.go:39 [It] should mirror a custom Endpoints resource through create update and delete [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: mirroring a new custom Endpoint Jun 3 11:27:38.616: INFO: Waiting for at least 1 EndpointSlice to exist, got 0 STEP: mirroring an update to a custom Endpoint Jun 3 11:27:40.678: INFO: Expected EndpointSlice to have 10.2.3.4 as address, got 10.1.2.3 STEP: mirroring deletion of a custom Endpoint Jun 3 11:27:42.691: INFO: Waiting for 0 EndpointSlices to exist, got 1 [AfterEach] [sig-network] EndpointSliceMirroring /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:44.696: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "endpointslicemirroring-2305" for this suite. • [SLOW TEST:6.138 seconds] [sig-network] EndpointSliceMirroring /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should mirror a custom Endpoints resource through create update and delete [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] EndpointSliceMirroring should mirror a custom Endpoints resource through create update and delete [Conformance]","total":-1,"completed":12,"skipped":214,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:41.234: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via environment variable [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap configmap-6496/configmap-test-4e5c5fbb-fc17-4085-93c4-f6eaf87c2b56 STEP: Creating a pod to test consume configMaps Jun 3 11:27:41.392: INFO: Waiting up to 5m0s for pod "pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e" in namespace "configmap-6496" to be "Succeeded or Failed" Jun 3 11:27:41.403: INFO: Pod "pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e": Phase="Pending", Reason="", readiness=false. Elapsed: 10.770427ms Jun 3 11:27:43.407: INFO: Pod "pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01504094s Jun 3 11:27:45.412: INFO: Pod "pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019292848s STEP: Saw pod success Jun 3 11:27:45.412: INFO: Pod "pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e" satisfied condition "Succeeded or Failed" Jun 3 11:27:45.415: INFO: Trying to get logs from node v1.21-worker2 pod pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e container env-test: STEP: delete the pod Jun 3 11:27:45.430: INFO: Waiting for pod pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e to disappear Jun 3 11:27:45.434: INFO: Pod pod-configmaps-23e02cc4-9940-43d2-9ac4-50ffe412987e no longer exists [AfterEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:45.435: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-6496" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","total":-1,"completed":10,"skipped":256,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:45.478: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should contain environment variables for services [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:27:45.521: INFO: The status of Pod server-envvars-0dbd12b6-e0a6-41e5-8fb4-82db862fd3d7 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:47.526: INFO: The status of Pod server-envvars-0dbd12b6-e0a6-41e5-8fb4-82db862fd3d7 is Running (Ready = true) Jun 3 11:27:47.544: INFO: Waiting up to 5m0s for pod "client-envvars-758a5609-5105-4f0f-86ee-900ac4e89568" in namespace "pods-3103" to be "Succeeded or Failed" Jun 3 11:27:47.550: INFO: Pod "client-envvars-758a5609-5105-4f0f-86ee-900ac4e89568": Phase="Pending", Reason="", readiness=false. Elapsed: 6.635703ms Jun 3 11:27:49.555: INFO: Pod "client-envvars-758a5609-5105-4f0f-86ee-900ac4e89568": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011328702s STEP: Saw pod success Jun 3 11:27:49.555: INFO: Pod "client-envvars-758a5609-5105-4f0f-86ee-900ac4e89568" satisfied condition "Succeeded or Failed" Jun 3 11:27:49.558: INFO: Trying to get logs from node v1.21-worker pod client-envvars-758a5609-5105-4f0f-86ee-900ac4e89568 container env3cont: STEP: delete the pod Jun 3 11:27:49.575: INFO: Waiting for pod client-envvars-758a5609-5105-4f0f-86ee-900ac4e89568 to disappear Jun 3 11:27:49.579: INFO: Pod client-envvars-758a5609-5105-4f0f-86ee-900ac4e89568 no longer exists [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:49.579: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-3103" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":-1,"completed":11,"skipped":271,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:44.720: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:69 [It] should create a PodDisruptionBudget [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pdb STEP: Waiting for the pdb to be processed STEP: updating the pdb STEP: Waiting for the pdb to be processed STEP: patching the pdb STEP: Waiting for the pdb to be processed STEP: Waiting for the pdb to be deleted [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:51.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-7214" for this suite. • [SLOW TEST:6.432 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should create a PodDisruptionBudget [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController should create a PodDisruptionBudget [Conformance]","total":-1,"completed":13,"skipped":220,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:49.624: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:38 [BeforeEach] when scheduling a busybox command that always fails in a pod /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:82 [It] should have an terminated reason [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:53.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-7033" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","total":-1,"completed":12,"skipped":288,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:25:51.355: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset W0603 11:25:51.389623 19 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+ Jun 3 11:25:51.389: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 3 11:25:51.396: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105 STEP: Creating service test in namespace statefulset-4625 [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Initializing watcher for selector baz=blah,foo=bar STEP: Creating stateful set ss in namespace statefulset-4625 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-4625 Jun 3 11:25:51.409: INFO: Found 0 stateful pods, waiting for 1 Jun 3 11:26:01.479: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod Jun 3 11:26:01.483: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:26:02.069: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:26:02.069: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:26:02.070: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:26:02.073: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Jun 3 11:26:12.081: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:26:12.081: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:26:12.188: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999498s Jun 3 11:26:13.194: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.995557621s Jun 3 11:26:14.198: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.990467903s Jun 3 11:26:15.489: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.985283823s Jun 3 11:26:16.883: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.694765579s Jun 3 11:26:17.886: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.301557157s Jun 3 11:26:18.890: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.298544822s Jun 3 11:26:19.894: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.293637455s Jun 3 11:26:20.897: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.290338087s Jun 3 11:26:21.901: INFO: Verifying statefulset ss doesn't scale past 1 for another 286.222846ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-4625 Jun 3 11:26:22.904: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:26:23.151: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 3 11:26:23.151: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:26:23.151: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:26:23.154: INFO: Found 1 stateful pods, waiting for 3 Jun 3 11:26:33.280: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:26:33.281: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:26:33.281: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Pending - Ready=false Jun 3 11:26:43.160: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:26:43.160: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:26:43.160: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Verifying that stateful set ss was scaled up in order STEP: Scale down will halt with unhealthy stateful pod Jun 3 11:26:43.167: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:26:43.372: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:26:43.372: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:26:43.372: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:26:43.372: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:26:43.585: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:26:43.585: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:26:43.585: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:26:43.586: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:26:43.814: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:26:43.814: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:26:43.814: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:26:43.814: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:26:43.817: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 Jun 3 11:26:53.826: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:26:53.827: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:26:53.827: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:26:53.839: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999565s Jun 3 11:26:54.883: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.995297033s Jun 3 11:26:55.888: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.951266107s Jun 3 11:26:56.893: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.946547434s Jun 3 11:26:57.897: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.941917963s Jun 3 11:26:58.901: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.938354491s Jun 3 11:26:59.906: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.933105749s Jun 3 11:27:00.910: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.928851394s Jun 3 11:27:01.915: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.923947711s Jun 3 11:27:02.919: INFO: Verifying statefulset ss doesn't scale past 3 for another 920.00637ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-4625 Jun 3 11:27:03.923: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:27:04.131: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 3 11:27:04.131: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:27:04.131: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:27:04.131: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:27:04.349: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 3 11:27:04.349: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:27:04.349: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:27:04.349: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-4625 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:27:04.596: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 3 11:27:04.596: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:27:04.596: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:27:04.596: INFO: Scaling statefulset ss to 0 STEP: Verifying that stateful set ss was scaled down in reverse order [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116 Jun 3 11:27:54.612: INFO: Deleting all statefulset in ns statefulset-4625 Jun 3 11:27:54.615: INFO: Scaling statefulset ss to 0 Jun 3 11:27:54.626: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:27:54.629: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:54.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-4625" for this suite. • [SLOW TEST:123.294 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95 Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","total":-1,"completed":1,"skipped":17,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:28.822: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service in namespace services-8508 STEP: creating service affinity-nodeport-transition in namespace services-8508 STEP: creating replication controller affinity-nodeport-transition in namespace services-8508 I0603 11:27:28.863067 22 runners.go:190] Created replication controller with name: affinity-nodeport-transition, namespace: services-8508, replica count: 3 I0603 11:27:31.914346 22 runners.go:190] affinity-nodeport-transition Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:27:34.915027 22 runners.go:190] affinity-nodeport-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:27:34.929: INFO: Creating new exec pod Jun 3 11:27:39.984: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-8508 exec execpod-affinitykzt79 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-transition 80' Jun 3 11:27:40.388: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-transition 80\nConnection to affinity-nodeport-transition 80 port [tcp/http] succeeded!\n" Jun 3 11:27:40.388: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:27:40.388: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-8508 exec execpod-affinitykzt79 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.8.93 80' Jun 3 11:27:40.729: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.8.93 80\nConnection to 10.96.8.93 80 port [tcp/http] succeeded!\n" Jun 3 11:27:40.729: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:27:40.729: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-8508 exec execpod-affinitykzt79 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.4 30891' Jun 3 11:27:41.294: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.4 30891\nConnection to 172.18.0.4 30891 port [tcp/*] succeeded!\n" Jun 3 11:27:41.294: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:27:41.294: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-8508 exec execpod-affinitykzt79 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.2 30891' Jun 3 11:27:41.508: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.2 30891\nConnection to 172.18.0.2 30891 port [tcp/*] succeeded!\n" Jun 3 11:27:41.508: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:27:41.686: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-8508 exec execpod-affinitykzt79 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.4:30891/ ; done' Jun 3 11:27:42.069: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n" Jun 3 11:27:42.069: INFO: stdout: "\naffinity-nodeport-transition-fjq5d\naffinity-nodeport-transition-fbwrz\naffinity-nodeport-transition-fbwrz\naffinity-nodeport-transition-fjq5d\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-fbwrz\naffinity-nodeport-transition-fjq5d\naffinity-nodeport-transition-fbwrz\naffinity-nodeport-transition-fjq5d\naffinity-nodeport-transition-fjq5d\naffinity-nodeport-transition-fjq5d\naffinity-nodeport-transition-fbwrz\naffinity-nodeport-transition-fbwrz\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-fjq5d" Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fjq5d Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fbwrz Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fbwrz Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fjq5d Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fbwrz Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fjq5d Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fbwrz Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fjq5d Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fjq5d Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fjq5d Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fbwrz Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fbwrz Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.069: INFO: Received response from host: affinity-nodeport-transition-fjq5d Jun 3 11:27:42.078: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-8508 exec execpod-affinitykzt79 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.4:30891/ ; done' Jun 3 11:27:42.550: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:30891/\n" Jun 3 11:27:42.550: INFO: stdout: "\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l\naffinity-nodeport-transition-g7j4l" Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Received response from host: affinity-nodeport-transition-g7j4l Jun 3 11:27:42.550: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-nodeport-transition in namespace services-8508, will wait for the garbage collector to delete the pods Jun 3 11:27:42.643: INFO: Deleting ReplicationController affinity-nodeport-transition took: 4.878332ms Jun 3 11:27:42.743: INFO: Terminating ReplicationController affinity-nodeport-transition pods took: 100.806108ms [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:55.277: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8508" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:26.465 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance]","total":-1,"completed":4,"skipped":104,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:55.315: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] getting/updating/patching custom resource definition status sub-resource works [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:27:55.401: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:55.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-1114" for this suite. • ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:54.702: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0644 on tmpfs Jun 3 11:27:54.738: INFO: Waiting up to 5m0s for pod "pod-cabd2683-1650-48c5-90c9-67179f34d170" in namespace "emptydir-7295" to be "Succeeded or Failed" Jun 3 11:27:54.741: INFO: Pod "pod-cabd2683-1650-48c5-90c9-67179f34d170": Phase="Pending", Reason="", readiness=false. Elapsed: 2.910107ms Jun 3 11:27:56.746: INFO: Pod "pod-cabd2683-1650-48c5-90c9-67179f34d170": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007730026s Jun 3 11:27:58.750: INFO: Pod "pod-cabd2683-1650-48c5-90c9-67179f34d170": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011160497s STEP: Saw pod success Jun 3 11:27:58.750: INFO: Pod "pod-cabd2683-1650-48c5-90c9-67179f34d170" satisfied condition "Succeeded or Failed" Jun 3 11:27:58.752: INFO: Trying to get logs from node v1.21-worker pod pod-cabd2683-1650-48c5-90c9-67179f34d170 container test-container: STEP: delete the pod Jun 3 11:27:58.765: INFO: Waiting for pod pod-cabd2683-1650-48c5-90c9-67179f34d170 to disappear Jun 3 11:27:58.768: INFO: Pod pod-cabd2683-1650-48c5-90c9-67179f34d170 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:27:58.768: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7295" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":44,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] KubeletManagedEtcHosts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:51.221: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts STEP: Waiting for a default service account to be provisioned in namespace [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Setting up the test STEP: Creating hostNetwork=false pod Jun 3 11:27:51.265: INFO: The status of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:53.272: INFO: The status of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:55.278: INFO: The status of Pod test-pod is Running (Ready = true) STEP: Creating hostNetwork=true pod Jun 3 11:27:55.290: INFO: The status of Pod test-host-network-pod is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:57.295: INFO: The status of Pod test-host-network-pod is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:59.296: INFO: The status of Pod test-host-network-pod is Running (Ready = true) STEP: Running the test STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false Jun 3 11:27:59.299: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:59.299: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:59.415: INFO: Exec stderr: "" Jun 3 11:27:59.415: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:59.415: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:59.525: INFO: Exec stderr: "" Jun 3 11:27:59.525: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:59.525: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:59.606: INFO: Exec stderr: "" Jun 3 11:27:59.606: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:59.606: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:59.723: INFO: Exec stderr: "" STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount Jun 3 11:27:59.723: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:59.723: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:59.829: INFO: Exec stderr: "" Jun 3 11:27:59.829: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:59.829: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:27:59.945: INFO: Exec stderr: "" STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true Jun 3 11:27:59.945: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:59.945: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:28:00.050: INFO: Exec stderr: "" Jun 3 11:28:00.050: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:28:00.050: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:28:00.159: INFO: Exec stderr: "" Jun 3 11:28:00.159: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:28:00.159: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:28:00.273: INFO: Exec stderr: "" Jun 3 11:28:00.273: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-263 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:28:00.273: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:28:00.385: INFO: Exec stderr: "" [AfterEach] [sig-node] KubeletManagedEtcHosts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:00.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-kubelet-etc-hosts-263" for this suite. • [SLOW TEST:9.172 seconds] [sig-node] KubeletManagedEtcHosts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":254,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:53.698: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should create services for rc [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating Agnhost RC Jun 3 11:27:53.733: INFO: namespace kubectl-6149 Jun 3 11:27:53.733: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-6149 create -f -' Jun 3 11:27:54.128: INFO: stderr: "" Jun 3 11:27:54.128: INFO: stdout: "replicationcontroller/agnhost-primary created\n" STEP: Waiting for Agnhost primary to start. Jun 3 11:27:55.134: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:27:55.134: INFO: Found 0 / 1 Jun 3 11:27:56.134: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:27:56.134: INFO: Found 1 / 1 Jun 3 11:27:56.134: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Jun 3 11:27:56.138: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:27:56.138: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Jun 3 11:27:56.138: INFO: wait on agnhost-primary startup in kubectl-6149 Jun 3 11:27:56.138: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-6149 logs agnhost-primary-pv6pk agnhost-primary' Jun 3 11:27:56.270: INFO: stderr: "" Jun 3 11:27:56.270: INFO: stdout: "Paused\n" STEP: exposing RC Jun 3 11:27:56.270: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-6149 expose rc agnhost-primary --name=rm2 --port=1234 --target-port=6379' Jun 3 11:27:56.422: INFO: stderr: "" Jun 3 11:27:56.422: INFO: stdout: "service/rm2 exposed\n" Jun 3 11:27:56.424: INFO: Service rm2 in namespace kubectl-6149 found. STEP: exposing service Jun 3 11:27:58.432: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-6149 expose service rm2 --name=rm3 --port=2345 --target-port=6379' Jun 3 11:27:58.576: INFO: stderr: "" Jun 3 11:27:58.576: INFO: stdout: "service/rm3 exposed\n" Jun 3 11:27:58.580: INFO: Service rm3 in namespace kubectl-6149 found. [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:00.587: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6149" for this suite. • [SLOW TEST:6.901 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl expose /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1223 should create services for rc [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance]","total":-1,"completed":13,"skipped":293,"failed":0} SSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:30.536: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:63 W0603 11:26:30.570190 32 warnings.go:70] batch/v1beta1 CronJob is deprecated in v1.21+, unavailable in v1.25+; use batch/v1 CronJob [It] should replace jobs when ReplaceConcurrent [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a ReplaceConcurrent cronjob STEP: Ensuring a job is scheduled STEP: Ensuring exactly one is scheduled STEP: Ensuring exactly one running job exists by listing jobs explicitly STEP: Ensuring the job is replaced with a new one STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:00.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-6073" for this suite. • [SLOW TEST:90.069 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should replace jobs when ReplaceConcurrent [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-apps] CronJob should replace jobs when ReplaceConcurrent [Conformance]","total":-1,"completed":6,"skipped":142,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:00.670: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should mount projected service account token [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test service account token: Jun 3 11:28:00.701: INFO: Waiting up to 5m0s for pod "test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1" in namespace "svcaccounts-2891" to be "Succeeded or Failed" Jun 3 11:28:00.703: INFO: Pod "test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.089222ms Jun 3 11:28:02.707: INFO: Pod "test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.005882701s Jun 3 11:28:04.711: INFO: Pod "test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010537265s Jun 3 11:28:06.715: INFO: Pod "test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013993105s STEP: Saw pod success Jun 3 11:28:06.715: INFO: Pod "test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1" satisfied condition "Succeeded or Failed" Jun 3 11:28:06.718: INFO: Trying to get logs from node v1.21-worker pod test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1 container agnhost-container: STEP: delete the pod Jun 3 11:28:06.733: INFO: Waiting for pod test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1 to disappear Jun 3 11:28:06.736: INFO: Pod test-pod-7d3252ee-5f07-4bbe-9b1a-7eb58cdf54f1 no longer exists [AfterEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:06.736: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-2891" for this suite. • [SLOW TEST:6.075 seconds] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should mount projected service account token [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts should mount projected service account token [Conformance]","total":-1,"completed":14,"skipped":337,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance]","total":-1,"completed":5,"skipped":117,"failed":0} [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:55.959: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should serve multiport endpoints from pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service multi-endpoint-test in namespace services-827 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-827 to expose endpoints map[] Jun 3 11:27:55.994: INFO: Failed go get Endpoints object: endpoints "multi-endpoint-test" not found Jun 3 11:27:57.004: INFO: successfully validated that service multi-endpoint-test in namespace services-827 exposes endpoints map[] STEP: Creating pod pod1 in namespace services-827 Jun 3 11:27:57.014: INFO: The status of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:59.018: INFO: The status of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:01.018: INFO: The status of Pod pod1 is Running (Ready = true) STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-827 to expose endpoints map[pod1:[100]] Jun 3 11:28:01.029: INFO: successfully validated that service multi-endpoint-test in namespace services-827 exposes endpoints map[pod1:[100]] STEP: Creating pod pod2 in namespace services-827 Jun 3 11:28:01.035: INFO: The status of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:03.039: INFO: The status of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:05.040: INFO: The status of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:07.039: INFO: The status of Pod pod2 is Running (Ready = true) STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-827 to expose endpoints map[pod1:[100] pod2:[101]] Jun 3 11:28:07.056: INFO: successfully validated that service multi-endpoint-test in namespace services-827 exposes endpoints map[pod1:[100] pod2:[101]] STEP: Deleting pod pod1 in namespace services-827 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-827 to expose endpoints map[pod2:[101]] Jun 3 11:28:07.075: INFO: successfully validated that service multi-endpoint-test in namespace services-827 exposes endpoints map[pod2:[101]] STEP: Deleting pod pod2 in namespace services-827 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-827 to expose endpoints map[] Jun 3 11:28:07.088: INFO: successfully validated that service multi-endpoint-test in namespace services-827 exposes endpoints map[] [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:07.097: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-827" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:11.145 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should serve multiport endpoints from pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods [Conformance]","total":-1,"completed":6,"skipped":117,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-instrumentation] Events /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:07.123: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a test event STEP: listing all events in all namespaces STEP: patching the test event STEP: fetching the test event STEP: deleting the test event STEP: listing all events in all namespaces [AfterEach] [sig-instrumentation] Events /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:07.181: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-6121" for this suite. • ------------------------------ {"msg":"PASSED [sig-instrumentation] Events should ensure that an event can be fetched, patched, deleted, and listed [Conformance]","total":-1,"completed":7,"skipped":129,"failed":0} SSS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:10.442: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should succeed in writing subpaths in container [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod STEP: waiting for pod running STEP: creating a file in subpath Jun 3 11:27:22.493: INFO: ExecWithOptions {Command:[/bin/sh -c touch /volume_mount/mypath/foo/test.log] Namespace:var-expansion-535 PodName:var-expansion-8ed943b5-8ec0-4d3b-8371-c14428284177 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:22.493: INFO: >>> kubeConfig: /root/.kube/config STEP: test for file in mounted path Jun 3 11:27:22.712: INFO: ExecWithOptions {Command:[/bin/sh -c test -f /subpath_mount/test.log] Namespace:var-expansion-535 PodName:var-expansion-8ed943b5-8ec0-4d3b-8371-c14428284177 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:27:22.712: INFO: >>> kubeConfig: /root/.kube/config STEP: updating the annotation value Jun 3 11:27:23.389: INFO: Successfully updated pod "var-expansion-8ed943b5-8ec0-4d3b-8371-c14428284177" STEP: waiting for annotated pod running STEP: deleting the pod gracefully Jun 3 11:27:23.393: INFO: Deleting pod "var-expansion-8ed943b5-8ec0-4d3b-8371-c14428284177" in namespace "var-expansion-535" Jun 3 11:27:23.400: INFO: Wait up to 5m0s for pod "var-expansion-8ed943b5-8ec0-4d3b-8371-c14428284177" to be fully deleted [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:09.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-535" for this suite. • [SLOW TEST:58.975 seconds] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should succeed in writing subpaths in container [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should succeed in writing subpaths in container [Slow] [Conformance]","total":-1,"completed":8,"skipped":94,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:58.806: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD with validation schema [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:27:58.835: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with known and required properties Jun 3 11:28:03.435: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 create -f -' Jun 3 11:28:03.906: INFO: stderr: "" Jun 3 11:28:03.906: INFO: stdout: "e2e-test-crd-publish-openapi-7737-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" Jun 3 11:28:03.907: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 delete e2e-test-crd-publish-openapi-7737-crds test-foo' Jun 3 11:28:04.028: INFO: stderr: "" Jun 3 11:28:04.028: INFO: stdout: "e2e-test-crd-publish-openapi-7737-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" Jun 3 11:28:04.028: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 apply -f -' Jun 3 11:28:04.352: INFO: stderr: "" Jun 3 11:28:04.352: INFO: stdout: "e2e-test-crd-publish-openapi-7737-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" Jun 3 11:28:04.352: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 delete e2e-test-crd-publish-openapi-7737-crds test-foo' Jun 3 11:28:04.475: INFO: stderr: "" Jun 3 11:28:04.475: INFO: stdout: "e2e-test-crd-publish-openapi-7737-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" STEP: client-side validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema Jun 3 11:28:04.475: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 create -f -' Jun 3 11:28:04.759: INFO: rc: 1 Jun 3 11:28:04.759: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 apply -f -' Jun 3 11:28:05.044: INFO: rc: 1 STEP: client-side validation (kubectl create and apply) rejects request without required properties Jun 3 11:28:05.044: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 create -f -' Jun 3 11:28:05.335: INFO: rc: 1 Jun 3 11:28:05.335: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 --namespace=crd-publish-openapi-4910 apply -f -' Jun 3 11:28:05.630: INFO: rc: 1 STEP: kubectl explain works to explain CR properties Jun 3 11:28:05.631: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 explain e2e-test-crd-publish-openapi-7737-crds' Jun 3 11:28:05.934: INFO: stderr: "" Jun 3 11:28:05.934: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7737-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" STEP: kubectl explain works to explain CR properties recursively Jun 3 11:28:05.935: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 explain e2e-test-crd-publish-openapi-7737-crds.metadata' Jun 3 11:28:06.222: INFO: stderr: "" Jun 3 11:28:06.222: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7737-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n clusterName\t\n The name of the cluster which the object belongs to. This is used to\n distinguish resources with same name and namespace in different clusters.\n This field is not set anywhere right now and apiserver is going to ignore\n it if set in create or update request.\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC.\n\n Populated by the system. Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested.\n\n Populated by the system when a graceful deletion is requested. Read-only.\n More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server.\n\n If this field is specified and the generated name exists, the server will\n NOT return a 409 - instead, it will either return 201 Created or 500 with\n Reason ServerTimeout indicating a unique name could not be found in the\n time allotted, and the client should retry (optionally after the time\n indicated in the Retry-After header).\n\n Applied only if Name is not specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within which each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty.\n\n Must be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources.\n\n Populated by the system. Read-only. Value must be treated as opaque by\n clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n SelfLink is a URL representing this object. Populated by the system.\n Read-only.\n\n DEPRECATED Kubernetes will stop propagating this field in 1.20 release and\n the field is planned to be removed in 1.21 release.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations.\n\n Populated by the system. Read-only. More info:\n http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" Jun 3 11:28:06.223: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 explain e2e-test-crd-publish-openapi-7737-crds.spec' Jun 3 11:28:06.513: INFO: stderr: "" Jun 3 11:28:06.513: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7737-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" Jun 3 11:28:06.513: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 explain e2e-test-crd-publish-openapi-7737-crds.spec.bars' Jun 3 11:28:06.806: INFO: stderr: "" Jun 3 11:28:06.806: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-7737-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n name\t -required-\n Name of Bar.\n\n" STEP: kubectl explain works to return error when explain is called on property that doesn't exist Jun 3 11:28:06.806: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4910 explain e2e-test-crd-publish-openapi-7737-crds.spec.bars2' Jun 3 11:28:07.104: INFO: rc: 1 [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:11.829: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-4910" for this suite. • [SLOW TEST:13.032 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD with validation schema [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":-1,"completed":3,"skipped":61,"failed":0} S ------------------------------ [BeforeEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:06.830: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not be blocked by dependency circle [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:06.881: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"55cfd1eb-566f-4a6d-80c5-92c9e3d5f7bf", Controller:(*bool)(0xc0053e398a), BlockOwnerDeletion:(*bool)(0xc0053e398b)}} Jun 3 11:28:06.889: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"4cd19f79-ed30-4d1c-964a-0c1eadae0b22", Controller:(*bool)(0xc000f249f2), BlockOwnerDeletion:(*bool)(0xc000f249f3)}} Jun 3 11:28:06.894: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"19ccc1ce-9ea7-4a87-9013-0f1bc29b6fa7", Controller:(*bool)(0xc0053e3bca), BlockOwnerDeletion:(*bool)(0xc0053e3bcb)}} [AfterEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:11.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-1459" for this suite. • [SLOW TEST:5.087 seconds] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not be blocked by dependency circle [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","total":-1,"completed":15,"skipped":381,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:26:59.755: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the rc1 STEP: create the rc2 STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well STEP: delete the rc simpletest-rc-to-be-deleted STEP: wait for the rc to be deleted STEP: Gathering metrics W0603 11:27:09.860774 17 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Jun 3 11:28:11.878: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. Jun 3 11:28:11.878: INFO: Deleting pod "simpletest-rc-to-be-deleted-785g9" in namespace "gc-3593" Jun 3 11:28:11.885: INFO: Deleting pod "simpletest-rc-to-be-deleted-872jh" in namespace "gc-3593" Jun 3 11:28:11.893: INFO: Deleting pod "simpletest-rc-to-be-deleted-96tjj" in namespace "gc-3593" Jun 3 11:28:11.900: INFO: Deleting pod "simpletest-rc-to-be-deleted-gv2l7" in namespace "gc-3593" Jun 3 11:28:11.907: INFO: Deleting pod "simpletest-rc-to-be-deleted-mkxt8" in namespace "gc-3593" [AfterEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:11.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-3593" for this suite. • [SLOW TEST:72.169 seconds] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","total":-1,"completed":7,"skipped":124,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:00.616: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:52 STEP: create the container to handle the HTTPGet hook request. Jun 3 11:28:00.650: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:02.654: INFO: The status of Pod pod-handle-http-request is Running (Ready = true) [It] should execute poststart exec hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the pod with lifecycle hook Jun 3 11:28:02.664: INFO: The status of Pod pod-with-poststart-exec-hook is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:04.672: INFO: The status of Pod pod-with-poststart-exec-hook is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:06.668: INFO: The status of Pod pod-with-poststart-exec-hook is Running (Ready = true) STEP: check poststart hook STEP: delete the pod with lifecycle hook Jun 3 11:28:06.683: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jun 3 11:28:06.687: INFO: Pod pod-with-poststart-exec-hook still exists Jun 3 11:28:08.688: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jun 3 11:28:08.692: INFO: Pod pod-with-poststart-exec-hook still exists Jun 3 11:28:10.688: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jun 3 11:28:10.691: INFO: Pod pod-with-poststart-exec-hook still exists Jun 3 11:28:12.688: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jun 3 11:28:12.692: INFO: Pod pod-with-poststart-exec-hook no longer exists [AfterEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:12.692: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-3462" for this suite. • [SLOW TEST:12.087 seconds] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:43 should execute poststart exec hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":147,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:01.973: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete pods created by rc when not orphaning [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the rc STEP: delete the rc STEP: wait for all pods to be garbage collected STEP: Gathering metrics W0603 11:27:12.027860 30 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Jun 3 11:28:14.046: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:14.046: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-6394" for this suite. • [SLOW TEST:72.084 seconds] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete pods created by rc when not orphaning [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","total":-1,"completed":3,"skipped":38,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:07.197: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:38 [It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:07.235: INFO: The status of Pod busybox-host-aliasesef23e403-86f3-4416-805d-12bfdb116bc7 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:09.239: INFO: The status of Pod busybox-host-aliasesef23e403-86f3-4416-805d-12bfdb116bc7 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:11.239: INFO: The status of Pod busybox-host-aliasesef23e403-86f3-4416-805d-12bfdb116bc7 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:13.278: INFO: The status of Pod busybox-host-aliasesef23e403-86f3-4416-805d-12bfdb116bc7 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:15.241: INFO: The status of Pod busybox-host-aliasesef23e403-86f3-4416-805d-12bfdb116bc7 is Running (Ready = true) [AfterEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:15.254: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-4211" for this suite. • [SLOW TEST:8.065 seconds] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 when scheduling a busybox Pod with hostAliases /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:137 should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":132,"failed":0} SSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:09.455: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating projection with secret that has name projected-secret-test-map-12b94e33-3f9c-4b00-9c1a-e630cd0e1154 STEP: Creating a pod to test consume secrets Jun 3 11:28:09.500: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109" in namespace "projected-4857" to be "Succeeded or Failed" Jun 3 11:28:09.502: INFO: Pod "pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109": Phase="Pending", Reason="", readiness=false. Elapsed: 2.550792ms Jun 3 11:28:11.507: INFO: Pod "pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007073348s Jun 3 11:28:13.510: INFO: Pod "pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010467629s Jun 3 11:28:15.515: INFO: Pod "pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015465976s Jun 3 11:28:17.579: INFO: Pod "pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.078915595s STEP: Saw pod success Jun 3 11:28:17.579: INFO: Pod "pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109" satisfied condition "Succeeded or Failed" Jun 3 11:28:17.681: INFO: Trying to get logs from node v1.21-worker pod pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109 container projected-secret-volume-test: STEP: delete the pod Jun 3 11:28:17.698: INFO: Waiting for pod pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109 to disappear Jun 3 11:28:17.701: INFO: Pod pod-projected-secrets-fcb05fd7-4254-4614-ad55-9ea047aab109 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:17.701: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4857" for this suite. • [SLOW TEST:8.257 seconds] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":115,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:17.774: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should check is all data is printed [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:17.838: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8813 version' Jun 3 11:28:17.939: INFO: stderr: "" Jun 3 11:28:17.939: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"21\", GitVersion:\"v1.21.1\", GitCommit:\"5e58841cce77d4bc13713ad2b91fa0d961e69192\", GitTreeState:\"clean\", BuildDate:\"2021-05-12T14:18:45Z\", GoVersion:\"go1.16.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"21\", GitVersion:\"v1.21.1\", GitCommit:\"5e58841cce77d4bc13713ad2b91fa0d961e69192\", GitTreeState:\"clean\", BuildDate:\"2021-05-18T01:10:20Z\", GoVersion:\"go1.16.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:17.940: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8813" for this suite. • ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance]","total":-1,"completed":10,"skipped":149,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:11.845: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] pod should support shared volumes between containers [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating Pod STEP: Reading file content from the nginx-container Jun 3 11:28:19.892: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-4586 PodName:pod-sharedvolume-79b7de1f-9aa7-43c2-8708-28f232a09f65 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:28:19.892: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:28:20.004: INFO: Exec stderr: "" [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:20.004: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4586" for this suite. • [SLOW TEST:8.169 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 pod should support shared volumes between containers [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":-1,"completed":4,"skipped":62,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:20.026: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should check if v1 is in available api versions [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: validating api versions Jun 3 11:28:20.062: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5318 api-versions' Jun 3 11:28:20.174: INFO: stderr: "" Jun 3 11:28:20.174: INFO: stdout: "admissionregistration.k8s.io/v1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ndiscovery.k8s.io/v1\ndiscovery.k8s.io/v1beta1\nevents.k8s.io/v1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nflowcontrol.apiserver.k8s.io/v1beta1\nk8s.cni.cncf.io/v1\nlitmuschaos.io/v1alpha1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1\nnode.k8s.io/v1beta1\npingcap.com/v1alpha1\npolicy/v1\npolicy/v1beta1\nprojectcontour.io/v1\nprojectcontour.io/v1alpha1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:20.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5318" for this suite. • ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance]","total":-1,"completed":5,"skipped":67,"failed":0} SSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:14.068: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:14.104: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes Jun 3 11:28:14.115: INFO: The status of Pod pod-exec-websocket-5e13d8c2-ee70-44f8-a542-c6f8baef2c52 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:16.119: INFO: The status of Pod pod-exec-websocket-5e13d8c2-ee70-44f8-a542-c6f8baef2c52 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:18.182: INFO: The status of Pod pod-exec-websocket-5e13d8c2-ee70-44f8-a542-c6f8baef2c52 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:20.119: INFO: The status of Pod pod-exec-websocket-5e13d8c2-ee70-44f8-a542-c6f8baef2c52 is Running (Ready = true) [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:20.239: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-3518" for this suite. • [SLOW TEST:6.222 seconds] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":42,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:20.189: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should patch a secret [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a secret STEP: listing secrets in all namespaces to ensure that there are more than zero STEP: patching the secret STEP: deleting the secret using a LabelSelector STEP: listing secrets in all namespaces, searching for label name and value in patch [AfterEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:20.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3416" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Secrets should patch a secret [Conformance]","total":-1,"completed":6,"skipped":70,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:15.281: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating secret secrets-738/secret-test-5ef13c33-4921-4e5b-8f3f-c4169fb63e9e STEP: Creating a pod to test consume secrets Jun 3 11:28:15.319: INFO: Waiting up to 5m0s for pod "pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da" in namespace "secrets-738" to be "Succeeded or Failed" Jun 3 11:28:15.322: INFO: Pod "pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da": Phase="Pending", Reason="", readiness=false. Elapsed: 3.145745ms Jun 3 11:28:17.326: INFO: Pod "pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007261897s Jun 3 11:28:19.331: INFO: Pod "pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012426321s Jun 3 11:28:21.335: INFO: Pod "pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016399745s STEP: Saw pod success Jun 3 11:28:21.336: INFO: Pod "pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da" satisfied condition "Succeeded or Failed" Jun 3 11:28:21.339: INFO: Trying to get logs from node v1.21-worker pod pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da container env-test: STEP: delete the pod Jun 3 11:28:21.355: INFO: Waiting for pod pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da to disappear Jun 3 11:28:21.358: INFO: Pod pod-configmaps-b8a644af-cb00-4f7c-a2bf-31a84fb9d4da no longer exists [AfterEach] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:21.358: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-738" for this suite. • [SLOW TEST:6.086 seconds] [sig-node] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should be consumable via the environment [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":143,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:11.956: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should serve a basic endpoint from pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service endpoint-test2 in namespace services-8522 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8522 to expose endpoints map[] Jun 3 11:28:11.990: INFO: Failed go get Endpoints object: endpoints "endpoint-test2" not found Jun 3 11:28:12.997: INFO: successfully validated that service endpoint-test2 in namespace services-8522 exposes endpoints map[] STEP: Creating pod pod1 in namespace services-8522 Jun 3 11:28:13.007: INFO: The status of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:15.011: INFO: The status of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:17.082: INFO: The status of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:19.012: INFO: The status of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:21.078: INFO: The status of Pod pod1 is Running (Ready = true) STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8522 to expose endpoints map[pod1:[80]] Jun 3 11:28:21.091: INFO: successfully validated that service endpoint-test2 in namespace services-8522 exposes endpoints map[pod1:[80]] STEP: Creating pod pod2 in namespace services-8522 Jun 3 11:28:21.099: INFO: The status of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:23.103: INFO: The status of Pod pod2 is Running (Ready = true) STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8522 to expose endpoints map[pod1:[80] pod2:[80]] Jun 3 11:28:23.117: INFO: successfully validated that service endpoint-test2 in namespace services-8522 exposes endpoints map[pod1:[80] pod2:[80]] STEP: Deleting pod pod1 in namespace services-8522 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8522 to expose endpoints map[pod2:[80]] Jun 3 11:28:23.133: INFO: successfully validated that service endpoint-test2 in namespace services-8522 exposes endpoints map[pod2:[80]] STEP: Deleting pod pod2 in namespace services-8522 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8522 to expose endpoints map[] Jun 3 11:28:23.147: INFO: successfully validated that service endpoint-test2 in namespace services-8522 exposes endpoints map[] [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:23.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8522" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:11.206 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should serve a basic endpoint from pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods [Conformance]","total":-1,"completed":8,"skipped":144,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:12.745: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:28:13.485: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:28:15.496: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:28:17.579: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:28:19.580: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316493, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:28:22.586: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] listing validating webhooks should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Listing all of the created validation webhooks STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Deleting the collection of validation webhooks STEP: Creating a configMap that does not comply to the validation webhook rules [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:23.754: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-3467" for this suite. STEP: Destroying namespace "webhook-3467-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:11.044 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 listing validating webhooks should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","total":-1,"completed":8,"skipped":167,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:00.426: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should run with the expected status [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpa': should get the expected 'State' STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpof': should get the expected 'State' STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpn': should get the expected 'State' STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] [AfterEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:27.027: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-5606" for this suite. • [SLOW TEST:26.608 seconds] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 blackbox test /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:41 when starting a container that exits /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:42 should run with the expected status [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","total":-1,"completed":15,"skipped":272,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:27.045: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should test the lifecycle of an Endpoint [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating an Endpoint STEP: waiting for available Endpoint STEP: listing all Endpoints STEP: updating the Endpoint STEP: fetching the Endpoint STEP: patching the Endpoint STEP: fetching the Endpoint STEP: deleting the Endpoint by Collection STEP: waiting for Endpoint deletion STEP: fetching the Endpoint [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:27.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3326" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • ------------------------------ {"msg":"PASSED [sig-network] Services should test the lifecycle of an Endpoint [Conformance]","total":-1,"completed":16,"skipped":277,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:21.390: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's args [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test substitution in container's args Jun 3 11:28:21.680: INFO: Waiting up to 5m0s for pod "var-expansion-19291462-755d-4b80-aab0-371ff53dee7b" in namespace "var-expansion-441" to be "Succeeded or Failed" Jun 3 11:28:21.685: INFO: Pod "var-expansion-19291462-755d-4b80-aab0-371ff53dee7b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.536168ms Jun 3 11:28:23.689: INFO: Pod "var-expansion-19291462-755d-4b80-aab0-371ff53dee7b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008933396s Jun 3 11:28:25.693: INFO: Pod "var-expansion-19291462-755d-4b80-aab0-371ff53dee7b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012648398s Jun 3 11:28:27.698: INFO: Pod "var-expansion-19291462-755d-4b80-aab0-371ff53dee7b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01770531s STEP: Saw pod success Jun 3 11:28:27.698: INFO: Pod "var-expansion-19291462-755d-4b80-aab0-371ff53dee7b" satisfied condition "Succeeded or Failed" Jun 3 11:28:27.702: INFO: Trying to get logs from node v1.21-worker pod var-expansion-19291462-755d-4b80-aab0-371ff53dee7b container dapi-container: STEP: delete the pod Jun 3 11:28:27.717: INFO: Waiting for pod var-expansion-19291462-755d-4b80-aab0-371ff53dee7b to disappear Jun 3 11:28:27.721: INFO: Pod var-expansion-19291462-755d-4b80-aab0-371ff53dee7b no longer exists [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:27.721: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-441" for this suite. • [SLOW TEST:6.341 seconds] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should allow substituting values in a container's args [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":-1,"completed":10,"skipped":153,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:27.765: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename certificates STEP: Waiting for a default service account to be provisioned in namespace [It] should support CSR API operations [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: getting /apis STEP: getting /apis/certificates.k8s.io STEP: getting /apis/certificates.k8s.io/v1 STEP: creating STEP: getting STEP: listing STEP: watching Jun 3 11:28:28.436: INFO: starting watch STEP: patching STEP: updating Jun 3 11:28:28.444: INFO: waiting for watch events with expected annotations Jun 3 11:28:28.444: INFO: saw patched and updated annotations STEP: getting /approval STEP: patching /approval STEP: updating /approval STEP: getting /status STEP: patching /status STEP: updating /status STEP: deleting STEP: deleting a collection [AfterEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:28.486: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "certificates-3463" for this suite. • ------------------------------ {"msg":"PASSED [sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance]","total":-1,"completed":11,"skipped":170,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:06.950: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name cm-test-opt-del-85cafe5b-5955-48d1-9832-e09ffbf2c574 STEP: Creating configMap with name cm-test-opt-upd-2f1ee759-c44e-4736-b6be-bd0492805dd0 STEP: Creating the pod Jun 3 11:27:07.008: INFO: The status of Pod pod-configmaps-da9a9843-7e02-4858-81c9-d24c10fea3f0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:09.012: INFO: The status of Pod pod-configmaps-da9a9843-7e02-4858-81c9-d24c10fea3f0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:11.013: INFO: The status of Pod pod-configmaps-da9a9843-7e02-4858-81c9-d24c10fea3f0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:13.013: INFO: The status of Pod pod-configmaps-da9a9843-7e02-4858-81c9-d24c10fea3f0 is Running (Ready = true) STEP: Deleting configmap cm-test-opt-del-85cafe5b-5955-48d1-9832-e09ffbf2c574 STEP: Updating configmap cm-test-opt-upd-2f1ee759-c44e-4736-b6be-bd0492805dd0 STEP: Creating configMap with name cm-test-opt-create-85756965-6d28-460d-8397-842ad2105b9c STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:29.736: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7511" for this suite. • [SLOW TEST:82.797 seconds] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":220,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:27.139: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should provide podname only [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:28:27.177: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8" in namespace "projected-2817" to be "Succeeded or Failed" Jun 3 11:28:27.179: INFO: Pod "downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8": Phase="Pending", Reason="", readiness=false. Elapsed: 1.996646ms Jun 3 11:28:29.184: INFO: Pod "downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007308099s Jun 3 11:28:31.190: INFO: Pod "downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012698506s STEP: Saw pod success Jun 3 11:28:31.190: INFO: Pod "downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8" satisfied condition "Succeeded or Failed" Jun 3 11:28:31.193: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8 container client-container: STEP: delete the pod Jun 3 11:28:31.209: INFO: Waiting for pod downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8 to disappear Jun 3 11:28:31.216: INFO: Pod downwardapi-volume-7936a744-1ac7-4342-bdb5-422013f19ea8 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:31.216: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2817" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","total":-1,"completed":17,"skipped":290,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:31.250: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should allow opting out of API token automount [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: getting the auto-created API token Jun 3 11:28:31.804: INFO: created pod pod-service-account-defaultsa Jun 3 11:28:31.804: INFO: pod pod-service-account-defaultsa service account token volume mount: true Jun 3 11:28:31.810: INFO: created pod pod-service-account-mountsa Jun 3 11:28:31.810: INFO: pod pod-service-account-mountsa service account token volume mount: true Jun 3 11:28:31.814: INFO: created pod pod-service-account-nomountsa Jun 3 11:28:31.814: INFO: pod pod-service-account-nomountsa service account token volume mount: false Jun 3 11:28:31.819: INFO: created pod pod-service-account-defaultsa-mountspec Jun 3 11:28:31.819: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true Jun 3 11:28:31.824: INFO: created pod pod-service-account-mountsa-mountspec Jun 3 11:28:31.824: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true Jun 3 11:28:31.828: INFO: created pod pod-service-account-nomountsa-mountspec Jun 3 11:28:31.828: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true Jun 3 11:28:31.831: INFO: created pod pod-service-account-defaultsa-nomountspec Jun 3 11:28:31.831: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false Jun 3 11:28:31.835: INFO: created pod pod-service-account-mountsa-nomountspec Jun 3 11:28:31.835: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false Jun 3 11:28:31.840: INFO: created pod pod-service-account-nomountsa-nomountspec Jun 3 11:28:31.840: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false [AfterEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:31.840: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-7501" for this suite. • ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]","total":-1,"completed":18,"skipped":300,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:29.849: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:86 [It] Deployment should have a working scale subresource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:29.887: INFO: Creating simple deployment test-new-deployment Jun 3 11:28:29.898: INFO: deployment "test-new-deployment" doesn't have the required revision set STEP: getting scale subresource STEP: updating a scale subresource STEP: verifying the deployment Spec.Replicas was modified STEP: Patch a scale subresource [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:80 Jun 3 11:28:31.953: INFO: Deployment "test-new-deployment": &Deployment{ObjectMeta:{test-new-deployment deployment-916 dfe69651-e262-42f7-a932-1456eb3a8ca9 3289186 3 2021-06-03 11:28:29 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 2021-06-03 11:28:29 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:28:31 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*4,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc005412058 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-06-03 11:28:31 +0000 UTC,LastTransitionTime:2021-06-03 11:28:31 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-new-deployment-847dcfb7fb" has successfully progressed.,LastUpdateTime:2021-06-03 11:28:31 +0000 UTC,LastTransitionTime:2021-06-03 11:28:29 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Jun 3 11:28:31.956: INFO: New ReplicaSet "test-new-deployment-847dcfb7fb" of Deployment "test-new-deployment": &ReplicaSet{ObjectMeta:{test-new-deployment-847dcfb7fb deployment-916 b732d67d-184e-459b-b8fb-48d0e5eb4207 3289188 2 2021-06-03 11:28:29 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-new-deployment dfe69651-e262-42f7-a932-1456eb3a8ca9 0xc005412617 0xc005412618}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:28:31 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"dfe69651-e262-42f7-a932-1456eb3a8ca9\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 847dcfb7fb,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0054126b8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:28:31.960: INFO: Pod "test-new-deployment-847dcfb7fb-2lp59" is available: &Pod{ObjectMeta:{test-new-deployment-847dcfb7fb-2lp59 test-new-deployment-847dcfb7fb- deployment-916 92250bba-444e-47ad-8df9-f359afcc0669 3289142 0 2021-06-03 11:28:29 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.243" ], "mac": "ea:74:c5:df:b7:2b", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.243" ], "mac": "ea:74:c5:df:b7:2b", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet test-new-deployment-847dcfb7fb b732d67d-184e-459b-b8fb-48d0e5eb4207 0xc0053bd257 0xc0053bd258}] [] [{kube-controller-manager Update v1 2021-06-03 11:28:29 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"b732d67d-184e-459b-b8fb-48d0e5eb4207\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:28:30 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:28:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.243\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-xtkkf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xtkkf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:28:29 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:28:31 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:28:31 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:28:29 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.4,PodIP:10.244.1.243,StartTime:2021-06-03 11:28:29 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:28:30 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://aaaa30b71af2ce3001308703d0dc21c7895c59f279875f18343d6bf94da72236,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.243,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:28:31.960: INFO: Pod "test-new-deployment-847dcfb7fb-fbrhl" is not available: &Pod{ObjectMeta:{test-new-deployment-847dcfb7fb-fbrhl test-new-deployment-847dcfb7fb- deployment-916 15931a8f-43d0-4cd0-ae9b-1fb4f937036f 3289193 0 2021-06-03 11:28:31 +0000 UTC map[name:httpd pod-template-hash:847dcfb7fb] map[] [{apps/v1 ReplicaSet test-new-deployment-847dcfb7fb b732d67d-184e-459b-b8fb-48d0e5eb4207 0xc0053bd497 0xc0053bd498}] [] [{kube-controller-manager Update v1 2021-06-03 11:28:31 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"b732d67d-184e-459b-b8fb-48d0e5eb4207\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vdj4c,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vdj4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:28:31 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:31.960: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-916" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] Deployment Deployment should have a working scale subresource [Conformance]","total":-1,"completed":8,"skipped":271,"failed":0} SSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:27:23.504: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name s-test-opt-del-553cabfc-6e41-4f03-b8e1-40e661d7dfa4 STEP: Creating secret with name s-test-opt-upd-54f69da0-72c4-49af-b5b3-e0aed65e24fd STEP: Creating the pod Jun 3 11:27:23.555: INFO: The status of Pod pod-projected-secrets-e618d8ae-4e63-43dc-af55-3a5d09a71b3a is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:25.560: INFO: The status of Pod pod-projected-secrets-e618d8ae-4e63-43dc-af55-3a5d09a71b3a is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:27.560: INFO: The status of Pod pod-projected-secrets-e618d8ae-4e63-43dc-af55-3a5d09a71b3a is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:29.560: INFO: The status of Pod pod-projected-secrets-e618d8ae-4e63-43dc-af55-3a5d09a71b3a is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:27:31.561: INFO: The status of Pod pod-projected-secrets-e618d8ae-4e63-43dc-af55-3a5d09a71b3a is Running (Ready = true) STEP: Deleting secret s-test-opt-del-553cabfc-6e41-4f03-b8e1-40e661d7dfa4 STEP: Updating secret s-test-opt-upd-54f69da0-72c4-49af-b5b3-e0aed65e24fd STEP: Creating secret with name s-test-opt-create-a687d69f-1ea9-43b7-8854-7486e45ec4af STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:32.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2639" for this suite. • [SLOW TEST:68.652 seconds] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":141,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:23.861: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a watch on configmaps with a certain label STEP: creating a new configmap STEP: modifying the configmap once STEP: changing the label value of the configmap STEP: Expecting to observe a delete notification for the watched object Jun 3 11:28:24.006: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-5847 dc93d7b0-605b-4910-b736-26463e9d4479 3288925 0 2021-06-03 11:28:23 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2021-06-03 11:28:23 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:28:24.006: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-5847 dc93d7b0-605b-4910-b736-26463e9d4479 3288926 0 2021-06-03 11:28:23 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2021-06-03 11:28:23 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:28:24.006: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-5847 dc93d7b0-605b-4910-b736-26463e9d4479 3288927 0 2021-06-03 11:28:23 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2021-06-03 11:28:23 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying the configmap a second time STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements STEP: changing the label value of the configmap back STEP: modifying the configmap a third time STEP: deleting the configmap STEP: Expecting to observe an add notification for the watched object when the label value was restored Jun 3 11:28:34.034: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-5847 dc93d7b0-605b-4910-b736-26463e9d4479 3289288 0 2021-06-03 11:28:23 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2021-06-03 11:28:23 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:28:34.034: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-5847 dc93d7b0-605b-4910-b736-26463e9d4479 3289289 0 2021-06-03 11:28:23 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2021-06-03 11:28:23 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:28:34.034: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-5847 dc93d7b0-605b-4910-b736-26463e9d4479 3289290 0 2021-06-03 11:28:23 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2021-06-03 11:28:23 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:34.035: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-5847" for this suite. • [SLOW TEST:10.183 seconds] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":-1,"completed":9,"skipped":210,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:31.989: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:28:32.019: INFO: Waiting up to 5m0s for pod "downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1" in namespace "downward-api-8353" to be "Succeeded or Failed" Jun 3 11:28:32.022: INFO: Pod "downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.323243ms Jun 3 11:28:34.025: INFO: Pod "downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006176816s Jun 3 11:28:36.030: INFO: Pod "downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010541616s STEP: Saw pod success Jun 3 11:28:36.030: INFO: Pod "downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1" satisfied condition "Succeeded or Failed" Jun 3 11:28:36.033: INFO: Trying to get logs from node v1.21-worker2 pod downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1 container client-container: STEP: delete the pod Jun 3 11:28:36.047: INFO: Waiting for pod downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1 to disappear Jun 3 11:28:36.050: INFO: Pod downwardapi-volume-498aa2ee-5ab7-4047-974a-ca38a32a38b1 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:36.050: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-8353" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":285,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:36.069: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should guarantee kube-root-ca.crt exist in any namespace [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:36.101: INFO: Got root ca configmap in namespace "svcaccounts-4381" Jun 3 11:28:36.105: INFO: Deleted root ca configmap in namespace "svcaccounts-4381" STEP: waiting for a new root ca configmap created Jun 3 11:28:36.609: INFO: Recreated root ca configmap in namespace "svcaccounts-4381" Jun 3 11:28:36.613: INFO: Updated root ca configmap in namespace "svcaccounts-4381" STEP: waiting for the root ca configmap reconciled Jun 3 11:28:37.119: INFO: Reconciled root ca configmap in namespace "svcaccounts-4381" [AfterEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:37.119: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-4381" for this suite. • ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts should guarantee kube-root-ca.crt exist in any namespace [Conformance]","total":-1,"completed":10,"skipped":289,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:31.873: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should update labels on modification [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating the pod Jun 3 11:28:31.911: INFO: The status of Pod labelsupdatea4469f80-c940-45a1-b9ba-ca83f4f527a4 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:33.915: INFO: The status of Pod labelsupdatea4469f80-c940-45a1-b9ba-ca83f4f527a4 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:28:35.916: INFO: The status of Pod labelsupdatea4469f80-c940-45a1-b9ba-ca83f4f527a4 is Running (Ready = true) Jun 3 11:28:36.439: INFO: Successfully updated pod "labelsupdatea4469f80-c940-45a1-b9ba-ca83f4f527a4" [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:38.453: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1783" for this suite. • [SLOW TEST:6.589 seconds] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should update labels on modification [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":-1,"completed":19,"skipped":313,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:34.079: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0777 on tmpfs Jun 3 11:28:34.115: INFO: Waiting up to 5m0s for pod "pod-b40209aa-ec5d-4e61-84fa-30f590423d72" in namespace "emptydir-7463" to be "Succeeded or Failed" Jun 3 11:28:34.118: INFO: Pod "pod-b40209aa-ec5d-4e61-84fa-30f590423d72": Phase="Pending", Reason="", readiness=false. Elapsed: 2.712183ms Jun 3 11:28:36.121: INFO: Pod "pod-b40209aa-ec5d-4e61-84fa-30f590423d72": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006646112s Jun 3 11:28:38.129: INFO: Pod "pod-b40209aa-ec5d-4e61-84fa-30f590423d72": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013791934s Jun 3 11:28:40.132: INFO: Pod "pod-b40209aa-ec5d-4e61-84fa-30f590423d72": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.017014433s STEP: Saw pod success Jun 3 11:28:40.132: INFO: Pod "pod-b40209aa-ec5d-4e61-84fa-30f590423d72" satisfied condition "Succeeded or Failed" Jun 3 11:28:40.135: INFO: Trying to get logs from node v1.21-worker pod pod-b40209aa-ec5d-4e61-84fa-30f590423d72 container test-container: STEP: delete the pod Jun 3 11:28:40.147: INFO: Waiting for pod pod-b40209aa-ec5d-4e61-84fa-30f590423d72 to disappear Jun 3 11:28:40.152: INFO: Pod pod-b40209aa-ec5d-4e61-84fa-30f590423d72 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:40.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7463" for this suite. • [SLOW TEST:6.082 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":10,"skipped":228,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:28.504: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a pod. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Pod that fits quota STEP: Ensuring ResourceQuota status captures the pod usage STEP: Not allowing a pod to be created that exceeds remaining quota STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) STEP: Ensuring a pod cannot update its resource requirements STEP: Ensuring attempts to update pod resource requirements did not change quota usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:41.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-5854" for this suite. • [SLOW TEST:13.115 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a pod. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":-1,"completed":12,"skipped":175,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:17.962: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with secret pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod pod-subpath-test-secret-m29l STEP: Creating a pod to test atomic-volume-subpath Jun 3 11:28:18.005: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-m29l" in namespace "subpath-2259" to be "Succeeded or Failed" Jun 3 11:28:18.008: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Pending", Reason="", readiness=false. Elapsed: 2.30383ms Jun 3 11:28:20.012: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 2.006830908s Jun 3 11:28:22.079: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 4.073902687s Jun 3 11:28:24.083: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 6.077904323s Jun 3 11:28:26.087: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 8.082016797s Jun 3 11:28:28.093: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 10.087588826s Jun 3 11:28:30.097: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 12.091651298s Jun 3 11:28:32.101: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 14.095923073s Jun 3 11:28:34.105: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 16.09962925s Jun 3 11:28:36.109: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 18.103558319s Jun 3 11:28:38.113: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 20.107955659s Jun 3 11:28:40.117: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Running", Reason="", readiness=true. Elapsed: 22.11170364s Jun 3 11:28:42.121: INFO: Pod "pod-subpath-test-secret-m29l": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.115312142s STEP: Saw pod success Jun 3 11:28:42.121: INFO: Pod "pod-subpath-test-secret-m29l" satisfied condition "Succeeded or Failed" Jun 3 11:28:42.124: INFO: Trying to get logs from node v1.21-worker2 pod pod-subpath-test-secret-m29l container test-container-subpath-secret-m29l: STEP: delete the pod Jun 3 11:28:42.139: INFO: Waiting for pod pod-subpath-test-secret-m29l to disappear Jun 3 11:28:42.142: INFO: Pod pod-subpath-test-secret-m29l no longer exists STEP: Deleting pod pod-subpath-test-secret-m29l Jun 3 11:28:42.142: INFO: Deleting pod "pod-subpath-test-secret-m29l" in namespace "subpath-2259" [AfterEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:42.144: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-2259" for this suite. • [SLOW TEST:24.190 seconds] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with secret pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]","total":-1,"completed":11,"skipped":155,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:37.138: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:28:37.396: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:28:39.408: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:28:41.413: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316517, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:28:44.418: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should include webhook resources in discovery documents [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: fetching the /apis discovery document STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document STEP: fetching the /apis/admissionregistration.k8s.io discovery document STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:44.430: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-541" for this suite. STEP: Destroying namespace "webhook-541-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:7.326 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should include webhook resources in discovery documents [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":-1,"completed":11,"skipped":296,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:11.932: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename endpointslice STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/endpointslice.go:49 [It] should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: referencing a single matching pod STEP: referencing matching pods with named port STEP: creating empty Endpoints and EndpointSlices for no matching Pods STEP: recreating EndpointSlices after they've been deleted Jun 3 11:28:37.030: INFO: EndpointSlice for Service endpointslice-5309/example-named-port not found [AfterEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:47.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "endpointslice-5309" for this suite. • [SLOW TEST:35.114 seconds] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] EndpointSlice should create Endpoints and EndpointSlices for Pods matching a Service [Conformance]","total":-1,"completed":16,"skipped":391,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:41.667: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-5b35c50c-6931-4ac9-9fec-8ed19d838a3d STEP: Creating a pod to test consume secrets Jun 3 11:28:41.805: INFO: Waiting up to 5m0s for pod "pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9" in namespace "secrets-6240" to be "Succeeded or Failed" Jun 3 11:28:41.807: INFO: Pod "pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.617082ms Jun 3 11:28:43.812: INFO: Pod "pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007036895s Jun 3 11:28:45.816: INFO: Pod "pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011598381s Jun 3 11:28:47.821: INFO: Pod "pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016528821s STEP: Saw pod success Jun 3 11:28:47.821: INFO: Pod "pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9" satisfied condition "Succeeded or Failed" Jun 3 11:28:47.825: INFO: Trying to get logs from node v1.21-worker pod pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9 container secret-volume-test: STEP: delete the pod Jun 3 11:28:47.841: INFO: Waiting for pod pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9 to disappear Jun 3 11:28:47.845: INFO: Pod pod-secrets-755dc201-6c29-4ec9-b187-831d0254edd9 no longer exists [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:47.845: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-6240" for this suite. • [SLOW TEST:6.187 seconds] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":13,"skipped":201,"failed":0} SS ------------------------------ [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:20.551: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod liveness-4988f217-447d-4240-9e37-be2f5b7a524c in namespace container-probe-5122 Jun 3 11:28:26.690: INFO: Started pod liveness-4988f217-447d-4240-9e37-be2f5b7a524c in namespace container-probe-5122 STEP: checking the pod's current state and verifying that restartCount is present Jun 3 11:28:26.693: INFO: Initial restart count of pod liveness-4988f217-447d-4240-9e37-be2f5b7a524c is 0 Jun 3 11:28:48.742: INFO: Restart count of pod container-probe-5122/liveness-4988f217-447d-4240-9e37-be2f5b7a524c is now 1 (22.048643636s elapsed) STEP: deleting the pod [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:48.751: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-5122" for this suite. • [SLOW TEST:28.209 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":83,"failed":0} SSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:23.204: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with projected pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod pod-subpath-test-projected-btjc STEP: Creating a pod to test atomic-volume-subpath Jun 3 11:28:23.243: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-btjc" in namespace "subpath-2528" to be "Succeeded or Failed" Jun 3 11:28:23.246: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.916258ms Jun 3 11:28:25.250: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007256334s Jun 3 11:28:27.254: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011133311s Jun 3 11:28:29.259: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 6.015550725s Jun 3 11:28:31.267: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 8.024042775s Jun 3 11:28:33.272: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 10.029116952s Jun 3 11:28:35.276: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 12.033156769s Jun 3 11:28:37.281: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 14.038221363s Jun 3 11:28:39.286: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 16.042889398s Jun 3 11:28:41.290: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 18.046704152s Jun 3 11:28:43.293: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 20.050151688s Jun 3 11:28:45.298: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 22.05451709s Jun 3 11:28:47.302: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Running", Reason="", readiness=true. Elapsed: 24.058628893s Jun 3 11:28:49.306: INFO: Pod "pod-subpath-test-projected-btjc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.062952811s STEP: Saw pod success Jun 3 11:28:49.306: INFO: Pod "pod-subpath-test-projected-btjc" satisfied condition "Succeeded or Failed" Jun 3 11:28:49.309: INFO: Trying to get logs from node v1.21-worker pod pod-subpath-test-projected-btjc container test-container-subpath-projected-btjc: STEP: delete the pod Jun 3 11:28:49.323: INFO: Waiting for pod pod-subpath-test-projected-btjc to disappear Jun 3 11:28:49.326: INFO: Pod pod-subpath-test-projected-btjc no longer exists STEP: Deleting pod pod-subpath-test-projected-btjc Jun 3 11:28:49.326: INFO: Deleting pod "pod-subpath-test-projected-btjc" in namespace "subpath-2528" [AfterEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:49.329: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-2528" for this suite. • [SLOW TEST:26.133 seconds] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with projected pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":-1,"completed":9,"skipped":171,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:38.474: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Service STEP: Creating a NodePort Service STEP: Not allowing a LoadBalancer Service with NodePort to be created that exceeds remaining quota STEP: Ensuring resource quota status captures service creation STEP: Deleting Services STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:49.583: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-1646" for this suite. • [SLOW TEST:11.116 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","total":-1,"completed":20,"skipped":319,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-scheduling] LimitRange /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:49.418: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename limitrange STEP: Waiting for a default service account to be provisioned in namespace [It] should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a LimitRange STEP: Setting up watch STEP: Submitting a LimitRange Jun 3 11:28:49.451: INFO: observed the limitRanges list STEP: Verifying LimitRange creation was observed STEP: Fetching the LimitRange to ensure it has proper values Jun 3 11:28:49.456: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] Jun 3 11:28:49.456: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Creating a Pod with no resource requirements STEP: Ensuring Pod has resource requirements applied from LimitRange Jun 3 11:28:49.461: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] Jun 3 11:28:49.461: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Creating a Pod with partial resource requirements STEP: Ensuring Pod has merged resource requirements applied from LimitRange Jun 3 11:28:49.472: INFO: Verifying requests: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] Jun 3 11:28:49.472: INFO: Verifying limits: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Failing to create a Pod with less than min resources STEP: Failing to create a Pod with more than max resources STEP: Updating a LimitRange STEP: Verifying LimitRange updating is effective STEP: Creating a Pod with less than former min resources STEP: Failing to create a Pod with more than max resources STEP: Deleting a LimitRange STEP: Verifying the LimitRange was deleted Jun 3 11:28:56.499: INFO: limitRange is already deleted STEP: Creating a Pod with more than former max resources [AfterEach] [sig-scheduling] LimitRange /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:56.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "limitrange-9340" for this suite. • [SLOW TEST:7.097 seconds] [sig-scheduling] LimitRange /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]","total":-1,"completed":10,"skipped":214,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:32.181: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] removes definition from spec when one version gets changed to not be served [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: set up a multi version CRD Jun 3 11:28:32.208: INFO: >>> kubeConfig: /root/.kube/config STEP: mark a version not serverd STEP: check the unserved version gets removed STEP: check the other version is not changed [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:56.598: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-7338" for this suite. • [SLOW TEST:24.426 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 removes definition from spec when one version gets changed to not be served [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":-1,"completed":8,"skipped":153,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] version v1 /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:48.780: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:48.803: INFO: Creating pod... Jun 3 11:28:48.814: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:49.818: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:50.823: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:51.819: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:52.820: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:53.818: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:54.819: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:55.819: INFO: Pod Quantity: 1 Status: Pending Jun 3 11:28:56.818: INFO: Pod Status: Running Jun 3 11:28:56.819: INFO: Creating service... Jun 3 11:28:56.826: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/pods/agnhost/proxy/some/path/with/DELETE Jun 3 11:28:56.830: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE Jun 3 11:28:56.830: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/pods/agnhost/proxy/some/path/with/GET Jun 3 11:28:56.834: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET Jun 3 11:28:56.834: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/pods/agnhost/proxy/some/path/with/HEAD Jun 3 11:28:56.838: INFO: http.Client request:HEAD | StatusCode:200 Jun 3 11:28:56.838: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/pods/agnhost/proxy/some/path/with/OPTIONS Jun 3 11:28:56.841: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS Jun 3 11:28:56.841: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/pods/agnhost/proxy/some/path/with/PATCH Jun 3 11:28:56.844: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH Jun 3 11:28:56.844: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/pods/agnhost/proxy/some/path/with/POST Jun 3 11:28:56.847: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST Jun 3 11:28:56.847: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/pods/agnhost/proxy/some/path/with/PUT Jun 3 11:28:56.851: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT Jun 3 11:28:56.851: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/services/test-service/proxy/some/path/with/DELETE Jun 3 11:28:56.855: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE Jun 3 11:28:56.855: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/services/test-service/proxy/some/path/with/GET Jun 3 11:28:56.859: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET Jun 3 11:28:56.859: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/services/test-service/proxy/some/path/with/HEAD Jun 3 11:28:56.863: INFO: http.Client request:HEAD | StatusCode:200 Jun 3 11:28:56.863: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/services/test-service/proxy/some/path/with/OPTIONS Jun 3 11:28:56.868: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS Jun 3 11:28:56.868: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/services/test-service/proxy/some/path/with/PATCH Jun 3 11:28:56.872: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH Jun 3 11:28:56.872: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/services/test-service/proxy/some/path/with/POST Jun 3 11:28:56.876: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST Jun 3 11:28:56.876: INFO: Starting http.Client for https://172.30.13.90:33295/api/v1/namespaces/proxy-5541/services/test-service/proxy/some/path/with/PUT Jun 3 11:28:56.881: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT [AfterEach] version v1 /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:56.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-5541" for this suite. • [SLOW TEST:8.110 seconds] [sig-network] Proxy /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 version v1 /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:74 A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Proxy version v1 A set of valid responses are returned for both pod and service ProxyWithPath [Conformance]","total":-1,"completed":8,"skipped":97,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:49.598: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-volume-map-3b0b89bb-96e2-45f9-a80c-dfb36a4a171f STEP: Creating a pod to test consume configMaps Jun 3 11:28:49.629: INFO: Waiting up to 5m0s for pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf" in namespace "configmap-8217" to be "Succeeded or Failed" Jun 3 11:28:49.631: INFO: Pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf": Phase="Pending", Reason="", readiness=false. Elapsed: 1.665591ms Jun 3 11:28:51.635: INFO: Pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006052137s Jun 3 11:28:53.639: INFO: Pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.009759802s Jun 3 11:28:55.643: INFO: Pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf": Phase="Pending", Reason="", readiness=false. Elapsed: 6.013889826s Jun 3 11:28:57.648: INFO: Pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf": Phase="Pending", Reason="", readiness=false. Elapsed: 8.018589692s Jun 3 11:28:59.652: INFO: Pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.022671247s STEP: Saw pod success Jun 3 11:28:59.652: INFO: Pod "pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf" satisfied condition "Succeeded or Failed" Jun 3 11:28:59.656: INFO: Trying to get logs from node v1.21-worker pod pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf container agnhost-container: STEP: delete the pod Jun 3 11:28:59.675: INFO: Waiting for pod pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf to disappear Jun 3 11:28:59.679: INFO: Pod pod-configmaps-57c54a32-b718-4cdf-ad1e-383df6ad3faf no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:28:59.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8217" for this suite. • [SLOW TEST:10.089 seconds] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":-1,"completed":21,"skipped":323,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:59.728: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] creating/deleting custom resource definition objects works [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:28:59.763: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:00.784: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-5211" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance]","total":-1,"completed":22,"skipped":344,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:44.478: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should verify ResourceQuota with best effort scope. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a ResourceQuota with best effort scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a ResourceQuota with not best effort scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a best-effort pod STEP: Ensuring resource quota with best effort scope captures the pod usage STEP: Ensuring resource quota with not best effort ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage STEP: Creating a not best-effort pod STEP: Ensuring resource quota with not best effort scope captures the pod usage STEP: Ensuring resource quota with best effort scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:00.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-5322" for this suite. • [SLOW TEST:16.602 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should verify ResourceQuota with best effort scope. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":-1,"completed":12,"skipped":303,"failed":0} SSS ------------------------------ [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:56.523: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should serve a basic image on each replica with a public image [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating replication controller my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a Jun 3 11:28:56.560: INFO: Pod name my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a: Found 0 pods out of 1 Jun 3 11:29:01.681: INFO: Pod name my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a: Found 1 pods out of 1 Jun 3 11:29:01.681: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a" are running Jun 3 11:29:01.684: INFO: Pod "my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a-nvcp8" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:28:56 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:28:57 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:28:57 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:28:56 +0000 UTC Reason: Message:}]) Jun 3 11:29:01.685: INFO: Trying to dial the pod Jun 3 11:29:06.985: INFO: Controller my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a: Got expected result from replica 1 [my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a-nvcp8]: "my-hostname-basic-db94d202-29b2-4541-a716-acf4471de41a-nvcp8", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:06.985: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-6115" for this suite. • [SLOW TEST:10.474 seconds] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance]","total":-1,"completed":11,"skipped":218,"failed":0} SS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:56.660: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:28:57.106: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:28:59.117: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:29:01.278: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:29:03.278: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316537, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:29:06.881: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:29:06.983: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-9919-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource that should be mutated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:10.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-8762" for this suite. STEP: Destroying namespace "webhook-8762-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:13.426 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":-1,"completed":9,"skipped":182,"failed":0} SSSS ------------------------------ [BeforeEach] version v1 /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:47.056: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy through a service and a pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: starting an echo server on multiple ports STEP: creating replication controller proxy-service-q7q87 in namespace proxy-8697 I0603 11:28:47.099685 34 runners.go:190] Created replication controller with name: proxy-service-q7q87, namespace: proxy-8697, replica count: 1 I0603 11:28:48.150924 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:49.151432 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:50.151549 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:51.151776 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:52.152040 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:53.152209 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:28:54.153406 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:28:55.153752 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:28:56.153968 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:28:57.154224 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:28:58.154470 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:28:59.154819 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:29:00.155271 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:29:01.155524 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:29:02.156311 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0603 11:29:03.157434 34 runners.go:190] proxy-service-q7q87 Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:29:03.578: INFO: setup took 16.491467504s, starting test cases STEP: running 16 cases, 20 attempts per case, 320 total attempts Jun 3 11:29:03.678: INFO: (0) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 100.648591ms) Jun 3 11:29:03.880: INFO: (0) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 301.789894ms) Jun 3 11:29:03.880: INFO: (0) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 301.902661ms) Jun 3 11:29:03.880: INFO: (0) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 302.431354ms) Jun 3 11:29:03.881: INFO: (0) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 302.51817ms) Jun 3 11:29:03.882: INFO: (0) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 303.55405ms) Jun 3 11:29:03.882: INFO: (0) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 304.601522ms) Jun 3 11:29:03.883: INFO: (0) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 304.697426ms) Jun 3 11:29:03.883: INFO: (0) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 304.827963ms) Jun 3 11:29:03.883: INFO: (0) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 305.0637ms) Jun 3 11:29:03.883: INFO: (0) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 305.231137ms) Jun 3 11:29:03.890: INFO: (0) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 311.792472ms) Jun 3 11:29:03.890: INFO: (0) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test<... (200; 588.982294ms) Jun 3 11:29:04.482: INFO: (1) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 589.167042ms) Jun 3 11:29:04.482: INFO: (1) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 589.082154ms) Jun 3 11:29:04.482: INFO: (1) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 589.243994ms) Jun 3 11:29:04.482: INFO: (1) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 589.494066ms) Jun 3 11:29:04.482: INFO: (1) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 589.536705ms) Jun 3 11:29:04.482: INFO: (1) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 589.541174ms) Jun 3 11:29:04.981: INFO: (1) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 1.088225021s) Jun 3 11:29:04.982: INFO: (1) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 1.089200251s) Jun 3 11:29:04.982: INFO: (1) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 1.089798182s) Jun 3 11:29:04.982: INFO: (1) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 1.089666277s) Jun 3 11:29:04.982: INFO: (1) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 1.089785472s) Jun 3 11:29:04.982: INFO: (1) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 1.089713116s) Jun 3 11:29:04.988: INFO: (2) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 5.743302ms) Jun 3 11:29:04.995: INFO: (2) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 12.129062ms) Jun 3 11:29:04.995: INFO: (2) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 12.22011ms) Jun 3 11:29:04.995: INFO: (2) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 12.396671ms) Jun 3 11:29:04.995: INFO: (2) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 12.442285ms) Jun 3 11:29:04.995: INFO: (2) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 12.745686ms) Jun 3 11:29:05.279: INFO: (2) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 296.956072ms) Jun 3 11:29:05.280: INFO: (2) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 297.083312ms) Jun 3 11:29:05.280: INFO: (2) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 297.272638ms) Jun 3 11:29:05.280: INFO: (2) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 297.34261ms) Jun 3 11:29:05.280: INFO: (2) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 297.658606ms) Jun 3 11:29:05.280: INFO: (2) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 297.67677ms) Jun 3 11:29:05.280: INFO: (2) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 297.843375ms) Jun 3 11:29:05.281: INFO: (2) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 298.092643ms) Jun 3 11:29:05.281: INFO: (2) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 401.016587ms) Jun 3 11:29:05.682: INFO: (3) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 401.544372ms) Jun 3 11:29:05.683: INFO: (3) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 402.058062ms) Jun 3 11:29:05.683: INFO: (3) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 402.023783ms) Jun 3 11:29:05.683: INFO: (3) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 402.102622ms) Jun 3 11:29:05.684: INFO: (3) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 402.512461ms) Jun 3 11:29:05.684: INFO: (3) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 402.593116ms) Jun 3 11:29:05.684: INFO: (3) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 402.59031ms) Jun 3 11:29:05.684: INFO: (3) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test<... (200; 403.121179ms) Jun 3 11:29:05.684: INFO: (3) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 403.203623ms) Jun 3 11:29:05.684: INFO: (3) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 403.370544ms) Jun 3 11:29:05.880: INFO: (4) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 195.228025ms) Jun 3 11:29:05.880: INFO: (4) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: ... (200; 195.857503ms) Jun 3 11:29:05.881: INFO: (4) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 195.818891ms) Jun 3 11:29:05.881: INFO: (4) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 195.980136ms) Jun 3 11:29:05.881: INFO: (4) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 196.415039ms) Jun 3 11:29:05.881: INFO: (4) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 196.692664ms) Jun 3 11:29:05.882: INFO: (4) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 196.975545ms) Jun 3 11:29:05.882: INFO: (4) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 196.825294ms) Jun 3 11:29:06.180: INFO: (4) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 495.09807ms) Jun 3 11:29:06.180: INFO: (4) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 495.033665ms) Jun 3 11:29:06.180: INFO: (4) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 495.561171ms) Jun 3 11:29:06.180: INFO: (4) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 495.526194ms) Jun 3 11:29:06.180: INFO: (4) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 495.685207ms) Jun 3 11:29:06.180: INFO: (4) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 495.680366ms) Jun 3 11:29:06.287: INFO: (5) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 106.678308ms) Jun 3 11:29:06.288: INFO: (5) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 107.181295ms) Jun 3 11:29:06.288: INFO: (5) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 107.2599ms) Jun 3 11:29:06.288: INFO: (5) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 107.455941ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 109.082566ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 108.951139ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 108.992467ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 109.102384ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 109.074491ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 109.088562ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 109.063786ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 109.08584ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 109.144171ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 109.210847ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 109.557107ms) Jun 3 11:29:06.290: INFO: (5) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: ... (200; 198.303654ms) Jun 3 11:29:06.489: INFO: (6) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 198.330148ms) Jun 3 11:29:06.489: INFO: (6) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 198.4089ms) Jun 3 11:29:06.489: INFO: (6) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 198.40798ms) Jun 3 11:29:06.489: INFO: (6) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 198.549545ms) Jun 3 11:29:06.489: INFO: (6) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 198.438243ms) Jun 3 11:29:06.489: INFO: (6) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: ... (200; 99.208641ms) Jun 3 11:29:06.589: INFO: (7) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 100.326215ms) Jun 3 11:29:06.590: INFO: (7) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 100.385269ms) Jun 3 11:29:06.590: INFO: (7) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 100.453305ms) Jun 3 11:29:06.590: INFO: (7) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 100.411131ms) Jun 3 11:29:06.590: INFO: (7) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 100.62577ms) Jun 3 11:29:06.590: INFO: (7) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 100.690398ms) Jun 3 11:29:06.678: INFO: (7) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 188.522787ms) Jun 3 11:29:06.682: INFO: (7) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 192.458833ms) Jun 3 11:29:06.682: INFO: (7) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 192.709121ms) Jun 3 11:29:06.682: INFO: (7) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 192.702362ms) Jun 3 11:29:06.682: INFO: (7) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 192.578316ms) Jun 3 11:29:06.682: INFO: (7) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 192.657699ms) Jun 3 11:29:06.687: INFO: (8) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 4.725091ms) Jun 3 11:29:06.687: INFO: (8) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 4.705327ms) Jun 3 11:29:06.687: INFO: (8) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 4.948096ms) Jun 3 11:29:06.687: INFO: (8) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 4.916129ms) Jun 3 11:29:06.786: INFO: (8) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: ... (200; 109.167199ms) Jun 3 11:29:06.792: INFO: (8) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 109.366276ms) Jun 3 11:29:06.792: INFO: (8) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 109.373077ms) Jun 3 11:29:06.792: INFO: (8) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 109.76528ms) Jun 3 11:29:06.792: INFO: (8) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 109.851579ms) Jun 3 11:29:06.882: INFO: (8) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 199.944294ms) Jun 3 11:29:06.983: INFO: (8) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 301.026388ms) Jun 3 11:29:06.983: INFO: (8) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 301.009736ms) Jun 3 11:29:06.990: INFO: (9) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 6.095255ms) Jun 3 11:29:06.990: INFO: (9) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 5.86365ms) Jun 3 11:29:06.990: INFO: (9) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: ... (200; 6.870097ms) Jun 3 11:29:06.991: INFO: (9) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 6.893218ms) Jun 3 11:29:06.990: INFO: (9) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 7.037694ms) Jun 3 11:29:06.991: INFO: (9) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 6.979814ms) Jun 3 11:29:06.991: INFO: (9) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 6.950667ms) Jun 3 11:29:06.991: INFO: (9) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 7.729305ms) Jun 3 11:29:06.992: INFO: (9) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 8.677295ms) Jun 3 11:29:06.994: INFO: (9) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 10.64283ms) Jun 3 11:29:06.995: INFO: (9) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 10.987692ms) Jun 3 11:29:06.995: INFO: (9) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 11.11397ms) Jun 3 11:29:06.995: INFO: (9) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 11.248269ms) Jun 3 11:29:06.995: INFO: (9) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 11.316738ms) Jun 3 11:29:06.998: INFO: (10) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 3.001757ms) Jun 3 11:29:06.998: INFO: (10) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 3.233533ms) Jun 3 11:29:07.001: INFO: (10) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 5.795076ms) Jun 3 11:29:07.001: INFO: (10) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 5.911745ms) Jun 3 11:29:07.001: INFO: (10) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 6.411564ms) Jun 3 11:29:07.002: INFO: (10) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 6.544461ms) Jun 3 11:29:07.002: INFO: (10) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 6.41767ms) Jun 3 11:29:07.002: INFO: (10) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 6.717661ms) Jun 3 11:29:07.002: INFO: (10) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 6.826298ms) Jun 3 11:29:07.003: INFO: (10) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 7.277214ms) Jun 3 11:29:07.003: INFO: (10) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 7.265038ms) Jun 3 11:29:07.003: INFO: (10) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 7.456639ms) Jun 3 11:29:07.003: INFO: (10) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 7.588684ms) Jun 3 11:29:07.006: INFO: (11) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.416695ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 3.722831ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 3.914321ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 4.063191ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 4.333336ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 4.314332ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 4.475139ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 4.451303ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 4.523146ms) Jun 3 11:29:07.007: INFO: (11) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 4.587459ms) Jun 3 11:29:07.008: INFO: (11) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 4.62902ms) Jun 3 11:29:07.008: INFO: (11) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 4.640105ms) Jun 3 11:29:07.008: INFO: (11) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 4.666397ms) Jun 3 11:29:07.008: INFO: (11) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 4.731515ms) Jun 3 11:29:07.008: INFO: (11) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 4.809751ms) Jun 3 11:29:07.011: INFO: (12) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 3.344526ms) Jun 3 11:29:07.011: INFO: (12) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 3.487337ms) Jun 3 11:29:07.012: INFO: (12) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: ... (200; 5.019891ms) Jun 3 11:29:07.013: INFO: (12) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 5.065041ms) Jun 3 11:29:07.013: INFO: (12) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 5.366482ms) Jun 3 11:29:07.013: INFO: (12) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 5.348427ms) Jun 3 11:29:07.013: INFO: (12) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 5.361939ms) Jun 3 11:29:07.014: INFO: (12) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 5.592216ms) Jun 3 11:29:07.014: INFO: (12) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 5.844629ms) Jun 3 11:29:07.014: INFO: (12) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 5.89132ms) Jun 3 11:29:07.017: INFO: (13) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 2.878715ms) Jun 3 11:29:07.017: INFO: (13) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 2.906744ms) Jun 3 11:29:07.017: INFO: (13) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 3.093197ms) Jun 3 11:29:07.017: INFO: (13) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.042043ms) Jun 3 11:29:07.017: INFO: (13) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 3.090777ms) Jun 3 11:29:07.017: INFO: (13) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 3.125578ms) Jun 3 11:29:07.018: INFO: (13) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 3.625508ms) Jun 3 11:29:07.018: INFO: (13) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test<... (200; 3.572282ms) Jun 3 11:29:07.023: INFO: (14) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 3.717787ms) Jun 3 11:29:07.023: INFO: (14) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 3.625741ms) Jun 3 11:29:07.023: INFO: (14) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 3.63644ms) Jun 3 11:29:07.023: INFO: (14) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.773893ms) Jun 3 11:29:07.023: INFO: (14) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 4.09723ms) Jun 3 11:29:07.023: INFO: (14) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 4.03476ms) Jun 3 11:29:07.023: INFO: (14) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 2.469676ms) Jun 3 11:29:07.027: INFO: (15) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.216439ms) Jun 3 11:29:07.027: INFO: (15) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 3.196173ms) Jun 3 11:29:07.027: INFO: (15) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.249468ms) Jun 3 11:29:07.027: INFO: (15) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 3.312264ms) Jun 3 11:29:07.027: INFO: (15) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test<... (200; 4.865703ms) Jun 3 11:29:07.029: INFO: (15) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 4.935763ms) Jun 3 11:29:07.029: INFO: (15) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 4.836025ms) Jun 3 11:29:07.038: INFO: (16) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 9.75488ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 10.00985ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 9.94308ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 9.954991ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 10.023965ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 10.003239ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 9.996633ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 9.989948ms) Jun 3 11:29:07.039: INFO: (16) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 10.54386ms) Jun 3 11:29:07.042: INFO: (17) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:1080/proxy/: test<... (200; 2.174607ms) Jun 3 11:29:07.043: INFO: (17) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.349461ms) Jun 3 11:29:07.043: INFO: (17) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 3.561889ms) Jun 3 11:29:07.043: INFO: (17) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 3.521498ms) Jun 3 11:29:07.043: INFO: (17) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 3.605831ms) Jun 3 11:29:07.043: INFO: (17) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 3.684742ms) Jun 3 11:29:07.043: INFO: (17) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 3.737447ms) Jun 3 11:29:07.043: INFO: (17) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test (200; 4.464165ms) Jun 3 11:29:07.046: INFO: (18) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 2.273389ms) Jun 3 11:29:07.046: INFO: (18) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 2.512992ms) Jun 3 11:29:07.047: INFO: (18) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 2.707354ms) Jun 3 11:29:07.047: INFO: (18) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 2.922542ms) Jun 3 11:29:07.047: INFO: (18) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 3.288044ms) Jun 3 11:29:07.047: INFO: (18) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.290804ms) Jun 3 11:29:07.047: INFO: (18) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 3.314007ms) Jun 3 11:29:07.048: INFO: (18) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 3.648298ms) Jun 3 11:29:07.048: INFO: (18) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 3.812186ms) Jun 3 11:29:07.048: INFO: (18) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 3.916136ms) Jun 3 11:29:07.048: INFO: (18) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 3.998606ms) Jun 3 11:29:07.048: INFO: (18) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 4.076221ms) Jun 3 11:29:07.048: INFO: (18) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test<... (200; 4.147034ms) Jun 3 11:29:07.048: INFO: (18) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 4.255339ms) Jun 3 11:29:07.049: INFO: (18) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 4.713414ms) Jun 3 11:29:07.051: INFO: (19) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:162/proxy/: bar (200; 2.143906ms) Jun 3 11:29:07.052: INFO: (19) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname2/proxy/: bar (200; 3.474906ms) Jun 3 11:29:07.052: INFO: (19) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname1/proxy/: tls baz (200; 3.596379ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname2/proxy/: bar (200; 3.651666ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/services/http:proxy-service-q7q87:portname1/proxy/: foo (200; 3.724543ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/services/https:proxy-service-q7q87:tlsportname2/proxy/: tls qux (200; 3.683033ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/services/proxy-service-q7q87:portname1/proxy/: foo (200; 3.875675ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:462/proxy/: tls qux (200; 3.724813ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:460/proxy/: tls baz (200; 3.881653ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:1080/proxy/: ... (200; 3.798289ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/pods/https:proxy-service-q7q87-dtjq7:443/proxy/: test<... (200; 4.02622ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 4.104916ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/pods/http:proxy-service-q7q87-dtjq7:160/proxy/: foo (200; 4.145984ms) Jun 3 11:29:07.053: INFO: (19) /api/v1/namespaces/proxy-8697/pods/proxy-service-q7q87-dtjq7/proxy/: test (200; 4.137424ms) STEP: deleting ReplicationController proxy-service-q7q87 in namespace proxy-8697, will wait for the garbage collector to delete the pods Jun 3 11:29:07.108: INFO: Deleting ReplicationController proxy-service-q7q87 took: 2.966462ms Jun 3 11:29:07.208: INFO: Terminating ReplicationController proxy-service-q7q87 pods took: 100.154652ms [AfterEach] version v1 /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:11.410: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-8697" for this suite. • [SLOW TEST:24.431 seconds] [sig-network] Proxy /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 version v1 /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:74 should proxy through a service and a pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance]","total":-1,"completed":17,"skipped":395,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:11.510: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:63 W0603 11:29:11.544537 34 warnings.go:70] batch/v1beta1 CronJob is deprecated in v1.21+, unavailable in v1.25+; use batch/v1 CronJob [It] should support CronJob API operations [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a cronjob STEP: creating STEP: getting STEP: listing STEP: watching Jun 3 11:29:11.555: INFO: starting watch STEP: cluster-wide listing STEP: cluster-wide watching Jun 3 11:29:11.560: INFO: starting watch STEP: patching STEP: updating Jun 3 11:29:11.577: INFO: waiting for watch events with expected annotations Jun 3 11:29:11.577: INFO: saw patched and updated annotations STEP: patching /status STEP: updating /status STEP: get /status STEP: deleting STEP: deleting a collection [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:11.614: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-1516" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] CronJob should support CronJob API operations [Conformance]","total":-1,"completed":18,"skipped":405,"failed":0} SSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:01.000: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should provide podname only [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:29:01.697: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798" in namespace "downward-api-4511" to be "Succeeded or Failed" Jun 3 11:29:01.699: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798": Phase="Pending", Reason="", readiness=false. Elapsed: 2.499269ms Jun 3 11:29:03.880: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798": Phase="Pending", Reason="", readiness=false. Elapsed: 2.183358177s Jun 3 11:29:06.179: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798": Phase="Pending", Reason="", readiness=false. Elapsed: 4.482151764s Jun 3 11:29:08.183: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798": Phase="Pending", Reason="", readiness=false. Elapsed: 6.486615739s Jun 3 11:29:10.187: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798": Phase="Pending", Reason="", readiness=false. Elapsed: 8.49007212s Jun 3 11:29:12.191: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798": Phase="Pending", Reason="", readiness=false. Elapsed: 10.494165526s Jun 3 11:29:14.195: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.498640259s STEP: Saw pod success Jun 3 11:29:14.196: INFO: Pod "downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798" satisfied condition "Succeeded or Failed" Jun 3 11:29:14.199: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798 container client-container: STEP: delete the pod Jun 3 11:29:14.215: INFO: Waiting for pod downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798 to disappear Jun 3 11:29:14.219: INFO: Pod downwardapi-volume-ada2a680-d27a-4ab8-b862-3ea903f27798 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:14.219: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-4511" for this suite. • [SLOW TEST:13.228 seconds] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should provide podname only [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","total":-1,"completed":23,"skipped":351,"failed":0} SSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:40.277: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service in namespace services-5538 STEP: creating service affinity-clusterip in namespace services-5538 STEP: creating replication controller affinity-clusterip in namespace services-5538 I0603 11:28:40.319472 32 runners.go:190] Created replication controller with name: affinity-clusterip, namespace: services-5538, replica count: 3 I0603 11:28:43.371061 32 runners.go:190] affinity-clusterip Pods: 3 out of 3 created, 1 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:46.371820 32 runners.go:190] affinity-clusterip Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:49.372294 32 runners.go:190] affinity-clusterip Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:28:49.377: INFO: Creating new exec pod Jun 3 11:29:00.389: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-5538 exec execpod-affinityr9cv9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip 80' Jun 3 11:29:00.933: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip 80\nConnection to affinity-clusterip 80 port [tcp/http] succeeded!\n" Jun 3 11:29:00.933: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:29:00.933: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-5538 exec execpod-affinityr9cv9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.200.60 80' Jun 3 11:29:01.594: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.200.60 80\nConnection to 10.96.200.60 80 port [tcp/http] succeeded!\n" Jun 3 11:29:01.594: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:29:01.594: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-5538 exec execpod-affinityr9cv9 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.96.200.60:80/ ; done' Jun 3 11:29:01.909: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.200.60:80/\n" Jun 3 11:29:01.909: INFO: stdout: "\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j\naffinity-clusterip-hq22j" Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.909: INFO: Received response from host: affinity-clusterip-hq22j Jun 3 11:29:01.910: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-clusterip in namespace services-5538, will wait for the garbage collector to delete the pods Jun 3 11:29:02.147: INFO: Deleting ReplicationController affinity-clusterip took: 4.078561ms Jun 3 11:29:02.248: INFO: Terminating ReplicationController affinity-clusterip pods took: 100.511064ms [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:15.558: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-5538" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:35.290 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance]","total":-1,"completed":11,"skipped":305,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:47.861: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should be able to change the type from NodePort to ExternalName [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a service nodeport-service with the type=NodePort in namespace services-6399 STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service STEP: creating service externalsvc in namespace services-6399 STEP: creating replication controller externalsvc in namespace services-6399 I0603 11:28:47.913608 22 runners.go:190] Created replication controller with name: externalsvc, namespace: services-6399, replica count: 2 I0603 11:28:50.965461 22 runners.go:190] externalsvc Pods: 2 out of 2 created, 1 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:53.965875 22 runners.go:190] externalsvc Pods: 2 out of 2 created, 1 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:28:56.966855 22 runners.go:190] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady STEP: changing the NodePort service to type=ExternalName Jun 3 11:28:56.982: INFO: Creating new exec pod Jun 3 11:29:05.280: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6399 exec execpodbd9f4 -- /bin/sh -x -c nslookup nodeport-service.services-6399.svc.cluster.local' Jun 3 11:29:06.305: INFO: stderr: "+ nslookup nodeport-service.services-6399.svc.cluster.local\n" Jun 3 11:29:06.305: INFO: stdout: "Server:\t\t10.96.0.10\nAddress:\t10.96.0.10#53\n\nnodeport-service.services-6399.svc.cluster.local\tcanonical name = externalsvc.services-6399.svc.cluster.local.\nName:\texternalsvc.services-6399.svc.cluster.local\nAddress: 10.96.185.58\n\n" STEP: deleting ReplicationController externalsvc in namespace services-6399, will wait for the garbage collector to delete the pods Jun 3 11:29:06.589: INFO: Deleting ReplicationController externalsvc took: 53.00974ms Jun 3 11:29:06.789: INFO: Terminating ReplicationController externalsvc pods took: 200.489755ms Jun 3 11:29:15.601: INFO: Cleaning up the NodePort to ExternalName test service [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:15.608: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-6399" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:27.765 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should be able to change the type from NodePort to ExternalName [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","total":-1,"completed":14,"skipped":203,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:56.957: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105 STEP: Creating service test in namespace statefulset-6213 [It] should have a working scale subresource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating statefulset ss in namespace statefulset-6213 Jun 3 11:28:56.993: INFO: Found 0 stateful pods, waiting for 1 Jun 3 11:29:06.998: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: getting scale subresource STEP: updating a scale subresource STEP: verifying the statefulset Spec.Replicas was modified STEP: Patch a scale subresource STEP: verifying the statefulset Spec.Replicas was modified [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116 Jun 3 11:29:07.016: INFO: Deleting all statefulset in ns statefulset-6213 Jun 3 11:29:07.018: INFO: Scaling statefulset ss to 0 Jun 3 11:29:17.032: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:29:17.035: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:17.051: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-6213" for this suite. • [SLOW TEST:20.101 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95 should have a working scale subresource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","total":-1,"completed":9,"skipped":137,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:15.614: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name projected-configmap-test-volume-map-9abfb074-93de-4fe2-8092-454077b525d4 STEP: Creating a pod to test consume configMaps Jun 3 11:29:15.670: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a" in namespace "projected-2045" to be "Succeeded or Failed" Jun 3 11:29:15.673: INFO: Pod "pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.709254ms Jun 3 11:29:17.678: INFO: Pod "pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007375588s Jun 3 11:29:19.682: INFO: Pod "pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011569169s STEP: Saw pod success Jun 3 11:29:19.682: INFO: Pod "pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a" satisfied condition "Succeeded or Failed" Jun 3 11:29:19.685: INFO: Trying to get logs from node v1.21-worker2 pod pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a container agnhost-container: STEP: delete the pod Jun 3 11:29:19.700: INFO: Waiting for pod pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a to disappear Jun 3 11:29:19.703: INFO: Pod pod-projected-configmaps-afc4d9e6-1e2c-4a1e-afb5-ee91b190900a no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:19.703: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2045" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":-1,"completed":12,"skipped":329,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] IngressClass API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:19.745: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename ingressclass STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] IngressClass API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/ingressclass.go:149 [It] should support creating IngressClass API operations [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: getting /apis STEP: getting /apis/networking.k8s.io STEP: getting /apis/networking.k8s.iov1 STEP: creating STEP: getting STEP: listing STEP: watching Jun 3 11:29:19.800: INFO: starting watch STEP: patching STEP: updating Jun 3 11:29:19.809: INFO: waiting for watch events with expected annotations Jun 3 11:29:19.809: INFO: saw patched and updated annotations STEP: deleting STEP: deleting a collection [AfterEach] [sig-network] IngressClass API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:19.830: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "ingressclass-4324" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] IngressClass API should support creating IngressClass API operations [Conformance]","total":-1,"completed":13,"skipped":344,"failed":0} SSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:01.088: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:69 [It] should observe PodDisruptionBudget status updated [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Waiting for the pdb to be processed STEP: Waiting for all pods to be running Jun 3 11:29:04.980: INFO: running pods: 0 < 3 Jun 3 11:29:06.985: INFO: running pods: 0 < 3 Jun 3 11:29:08.986: INFO: running pods: 1 < 3 Jun 3 11:29:10.986: INFO: running pods: 1 < 3 Jun 3 11:29:12.986: INFO: running pods: 2 < 3 Jun 3 11:29:14.984: INFO: running pods: 2 < 3 Jun 3 11:29:16.985: INFO: running pods: 2 < 3 Jun 3 11:29:18.985: INFO: running pods: 2 < 3 [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:20.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-2409" for this suite. • [SLOW TEST:19.907 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should observe PodDisruptionBudget status updated [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController should observe PodDisruptionBudget status updated [Conformance]","total":-1,"completed":13,"skipped":306,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:19.845: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0666 on tmpfs Jun 3 11:29:19.880: INFO: Waiting up to 5m0s for pod "pod-df3d9a96-3288-4a31-b396-f17288ea61e5" in namespace "emptydir-9728" to be "Succeeded or Failed" Jun 3 11:29:19.883: INFO: Pod "pod-df3d9a96-3288-4a31-b396-f17288ea61e5": Phase="Pending", Reason="", readiness=false. Elapsed: 3.01487ms Jun 3 11:29:21.888: INFO: Pod "pod-df3d9a96-3288-4a31-b396-f17288ea61e5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007215531s STEP: Saw pod success Jun 3 11:29:21.888: INFO: Pod "pod-df3d9a96-3288-4a31-b396-f17288ea61e5" satisfied condition "Succeeded or Failed" Jun 3 11:29:21.891: INFO: Trying to get logs from node v1.21-worker2 pod pod-df3d9a96-3288-4a31-b396-f17288ea61e5 container test-container: STEP: delete the pod Jun 3 11:29:21.906: INFO: Waiting for pod pod-df3d9a96-3288-4a31-b396-f17288ea61e5 to disappear Jun 3 11:29:21.909: INFO: Pod pod-df3d9a96-3288-4a31-b396-f17288ea61e5 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:21.909: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-9728" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":347,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:14.237: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:126 STEP: Setting up server cert STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication STEP: Deploying the custom resource conversion webhook pod STEP: Wait for the deployment to be ready Jun 3 11:29:14.797: INFO: new replicaset for deployment "sample-crd-conversion-webhook-deployment" is yet to be created Jun 3 11:29:16.807: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316554, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316554, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316554, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316554, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-697cdbd8f4\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:29:19.821: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 [It] should be able to convert a non homogeneous list of CRs [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:29:19.824: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating a v1 custom resource STEP: Create a v2 custom resource STEP: List CRs in v1 STEP: List CRs in v2 [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:23.041: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-webhook-3736" for this suite. [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:137 • [SLOW TEST:8.836 seconds] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to convert a non homogeneous list of CRs [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","total":-1,"completed":24,"skipped":354,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:21.042: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name projected-configmap-test-volume-map-1d8054c9-54df-4698-906c-7f2a2bcde3f8 STEP: Creating a pod to test consume configMaps Jun 3 11:29:21.083: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-dfbd8224-72f1-48c4-bbe5-0198d54ec2e8" in namespace "projected-6866" to be "Succeeded or Failed" Jun 3 11:29:21.085: INFO: Pod "pod-projected-configmaps-dfbd8224-72f1-48c4-bbe5-0198d54ec2e8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.277201ms Jun 3 11:29:23.088: INFO: Pod "pod-projected-configmaps-dfbd8224-72f1-48c4-bbe5-0198d54ec2e8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.005267813s STEP: Saw pod success Jun 3 11:29:23.088: INFO: Pod "pod-projected-configmaps-dfbd8224-72f1-48c4-bbe5-0198d54ec2e8" satisfied condition "Succeeded or Failed" Jun 3 11:29:23.090: INFO: Trying to get logs from node v1.21-worker2 pod pod-projected-configmaps-dfbd8224-72f1-48c4-bbe5-0198d54ec2e8 container agnhost-container: STEP: delete the pod Jun 3 11:29:23.102: INFO: Waiting for pod pod-projected-configmaps-dfbd8224-72f1-48c4-bbe5-0198d54ec2e8 to disappear Jun 3 11:29:23.104: INFO: Pod pod-projected-configmaps-dfbd8224-72f1-48c4-bbe5-0198d54ec2e8 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:23.104: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6866" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":334,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:23.091: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should update labels on modification [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating the pod Jun 3 11:29:23.127: INFO: The status of Pod labelsupdateedae43bc-7ada-48b4-8f39-82debd114edd is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:25.281: INFO: The status of Pod labelsupdateedae43bc-7ada-48b4-8f39-82debd114edd is Running (Ready = true) Jun 3 11:29:25.997: INFO: Successfully updated pod "labelsupdateedae43bc-7ada-48b4-8f39-82debd114edd" [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:28.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-3262" for this suite. • ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:21.959: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD preserving unknown fields at the schema root [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:29:21.989: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Jun 3 11:29:28.152: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-7967 --namespace=crd-publish-openapi-7967 create -f -' Jun 3 11:29:28.646: INFO: stderr: "" Jun 3 11:29:28.646: INFO: stdout: "e2e-test-crd-publish-openapi-8289-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" Jun 3 11:29:28.646: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-7967 --namespace=crd-publish-openapi-7967 delete e2e-test-crd-publish-openapi-8289-crds test-cr' Jun 3 11:29:28.769: INFO: stderr: "" Jun 3 11:29:28.770: INFO: stdout: "e2e-test-crd-publish-openapi-8289-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" Jun 3 11:29:28.770: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-7967 --namespace=crd-publish-openapi-7967 apply -f -' Jun 3 11:29:29.082: INFO: stderr: "" Jun 3 11:29:29.083: INFO: stdout: "e2e-test-crd-publish-openapi-8289-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" Jun 3 11:29:29.083: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-7967 --namespace=crd-publish-openapi-7967 delete e2e-test-crd-publish-openapi-8289-crds test-cr' Jun 3 11:29:29.199: INFO: stderr: "" Jun 3 11:29:29.199: INFO: stdout: "e2e-test-crd-publish-openapi-8289-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR Jun 3 11:29:29.199: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-7967 explain e2e-test-crd-publish-openapi-8289-crds' Jun 3 11:29:29.494: INFO: stderr: "" Jun 3 11:29:29.494: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-8289-crd\nVERSION: crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n \n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:34.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-7967" for this suite. • [SLOW TEST:12.238 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD preserving unknown fields at the schema root [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","total":-1,"completed":15,"skipped":368,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:34.216: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod STEP: submitting the pod to kubernetes Jun 3 11:29:34.260: INFO: The status of Pod pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:36.265: INFO: The status of Pod pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac is Running (Ready = true) STEP: verifying the pod is in kubernetes STEP: updating the pod Jun 3 11:29:36.786: INFO: Successfully updated pod "pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac" Jun 3 11:29:36.786: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac" in namespace "pods-6813" to be "terminated due to deadline exceeded" Jun 3 11:29:36.790: INFO: Pod "pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac": Phase="Running", Reason="", readiness=true. Elapsed: 3.763979ms Jun 3 11:29:38.794: INFO: Pod "pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.008336891s Jun 3 11:29:40.799: INFO: Pod "pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 4.012955068s Jun 3 11:29:40.799: INFO: Pod "pod-update-activedeadlineseconds-7285c07e-7d48-4ec6-8a48-d4764918a7ac" satisfied condition "terminated due to deadline exceeded" [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:40.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-6813" for this suite. • [SLOW TEST:6.593 seconds] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","total":-1,"completed":16,"skipped":377,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":-1,"completed":25,"skipped":364,"failed":0} [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:28.022: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:86 [It] deployment should support rollover [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:29:28.063: INFO: Pod name rollover-pod: Found 0 pods out of 1 Jun 3 11:29:33.067: INFO: Pod name rollover-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Jun 3 11:29:33.067: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready Jun 3 11:29:35.071: INFO: Creating deployment "test-rollover-deployment" Jun 3 11:29:35.078: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations Jun 3 11:29:37.084: INFO: Check revision of new replica set for deployment "test-rollover-deployment" Jun 3 11:29:37.091: INFO: Ensure that both replica sets have 1 created replica Jun 3 11:29:37.098: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update Jun 3 11:29:37.108: INFO: Updating deployment test-rollover-deployment Jun 3 11:29:37.108: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller Jun 3 11:29:39.115: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 Jun 3 11:29:39.121: INFO: Make sure deployment "test-rollover-deployment" is complete Jun 3 11:29:39.127: INFO: all replica sets need to contain the pod-template-hash label Jun 3 11:29:39.127: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316578, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-98c5f4599\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:29:41.184: INFO: all replica sets need to contain the pod-template-hash label Jun 3 11:29:41.184: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316578, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-98c5f4599\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:29:43.136: INFO: all replica sets need to contain the pod-template-hash label Jun 3 11:29:43.136: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316578, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-98c5f4599\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:29:45.184: INFO: all replica sets need to contain the pod-template-hash label Jun 3 11:29:45.184: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316578, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-98c5f4599\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:29:47.135: INFO: all replica sets need to contain the pod-template-hash label Jun 3 11:29:47.135: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316578, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316575, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-98c5f4599\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:29:49.136: INFO: Jun 3 11:29:49.136: INFO: Ensure that both old replica sets have no replicas [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:80 Jun 3 11:29:49.145: INFO: Deployment "test-rollover-deployment": &Deployment{ObjectMeta:{test-rollover-deployment deployment-2468 75dce929-ca1a-4ad6-9b36-1bb53791f208 3291619 2 2021-06-03 11:29:35 +0000 UTC map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2021-06-03 11:29:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:minReadySeconds":{},"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:29:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.32 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004eb6688 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-06-03 11:29:35 +0000 UTC,LastTransitionTime:2021-06-03 11:29:35 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-98c5f4599" has successfully progressed.,LastUpdateTime:2021-06-03 11:29:49 +0000 UTC,LastTransitionTime:2021-06-03 11:29:35 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Jun 3 11:29:49.149: INFO: New ReplicaSet "test-rollover-deployment-98c5f4599" of Deployment "test-rollover-deployment": &ReplicaSet{ObjectMeta:{test-rollover-deployment-98c5f4599 deployment-2468 abef51c0-0028-45cc-82d3-75c017fe4faa 3291608 2 2021-06-03 11:29:37 +0000 UTC map[name:rollover-pod pod-template-hash:98c5f4599] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment 75dce929-ca1a-4ad6-9b36-1bb53791f208 0xc000554f60 0xc000554f61}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:29:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"75dce929-ca1a-4ad6-9b36-1bb53791f208\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 98c5f4599,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:98c5f4599] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.32 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc000555068 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:29:49.149: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": Jun 3 11:29:49.150: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller deployment-2468 9fad7425-43e0-4e85-b37a-6fed45513fda 3291617 2 2021-06-03 11:29:28 +0000 UTC map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment 75dce929-ca1a-4ad6-9b36-1bb53791f208 0xc000554b1f 0xc000554b30}] [] [{e2e.test Update apps/v1 2021-06-03 11:29:28 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:29:49 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"75dce929-ca1a-4ad6-9b36-1bb53791f208\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod:httpd] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc000554c68 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:29:49.150: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-78bc8b888c deployment-2468 cce0ac39-9c6f-451b-a156-96e8b591d685 3291349 2 2021-06-03 11:29:35 +0000 UTC map[name:rollover-pod pod-template-hash:78bc8b888c] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment 75dce929-ca1a-4ad6-9b36-1bb53791f208 0xc000554d07 0xc000554d08}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:29:37 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"75dce929-ca1a-4ad6-9b36-1bb53791f208\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"redis-slave\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 78bc8b888c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:78bc8b888c] map[] [] [] []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc000554e28 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:29:49.154: INFO: Pod "test-rollover-deployment-98c5f4599-m78w4" is available: &Pod{ObjectMeta:{test-rollover-deployment-98c5f4599-m78w4 test-rollover-deployment-98c5f4599- deployment-2468 f4352f35-564b-450c-9814-f607b0711800 3291372 0 2021-06-03 11:29:37 +0000 UTC map[name:rollover-pod pod-template-hash:98c5f4599] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.18" ], "mac": "c2:fb:d5:b8:84:0d", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.18" ], "mac": "c2:fb:d5:b8:84:0d", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet test-rollover-deployment-98c5f4599 abef51c0-0028-45cc-82d3-75c017fe4faa 0xc000555bd0 0xc000555bd1}] [] [{kube-controller-manager Update v1 2021-06-03 11:29:37 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"abef51c0-0028-45cc-82d3-75c017fe4faa\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:29:37 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:29:38 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.18\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-nhvts,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:k8s.gcr.io/e2e-test-images/agnhost:2.32,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nhvts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:29:37 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:29:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:29:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:29:37 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.4,PodIP:10.244.1.18,StartTime:2021-06-03 11:29:37 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:29:38 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/agnhost:2.32,ImageID:k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1,ContainerID:containerd://cd1635ac4315954f65eab3bed17de4f6109a7d94b1de8da38b1c0130ef3d5b78,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.18,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:49.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-2468" for this suite. • [SLOW TEST:21.142 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support rollover [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","total":-1,"completed":26,"skipped":364,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:49.185: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:29:49.384: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes Jun 3 11:29:49.396: INFO: The status of Pod pod-logs-websocket-b787d392-50db-4b4d-bbab-dd26dd3e76c6 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:51.401: INFO: The status of Pod pod-logs-websocket-b787d392-50db-4b4d-bbab-dd26dd3e76c6 is Running (Ready = true) [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:51.430: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-6030" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":-1,"completed":27,"skipped":373,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:51.489: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] binary data should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-upd-ad297901-f010-4f5a-bf3f-434977649d5d STEP: Creating the pod STEP: Waiting for pod with text data STEP: Waiting for pod with binary data [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:53.556: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-6990" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":28,"skipped":378,"failed":0} SS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:07.005: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a job [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: delete a job STEP: deleting Job.batch foo in namespace job-4089, will wait for the garbage collector to delete the pods Jun 3 11:29:15.093: INFO: Deleting Job.batch foo took: 4.009018ms Jun 3 11:29:15.194: INFO: Terminating Job.batch foo pods took: 100.89095ms STEP: Ensuring job was deleted [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:55.097: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-4089" for this suite. • [SLOW TEST:48.102 seconds] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should delete a job [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","total":-1,"completed":12,"skipped":220,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:53.572: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:38 [It] should print the output to logs [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:29:53.613: INFO: The status of Pod busybox-scheduling-945cbc0e-17e6-45b9-ac6f-9b5e2360d993 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:55.618: INFO: The status of Pod busybox-scheduling-945cbc0e-17e6-45b9-ac6f-9b5e2360d993 is Running (Ready = true) [AfterEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:55.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-1618" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","total":-1,"completed":29,"skipped":380,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:23.122: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for services [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6099.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6099.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-6099.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-6099.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6099.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 43.97.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.97.43_udp@PTR;check="$$(dig +tcp +noall +answer +search 43.97.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.97.43_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6099.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6099.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-6099.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6099.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-6099.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6099.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 43.97.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.97.43_udp@PTR;check="$$(dig +tcp +noall +answer +search 43.97.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.97.43_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:29:25.483: INFO: Unable to read wheezy_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.486: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.489: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.492: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.511: INFO: Unable to read jessie_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.513: INFO: Unable to read jessie_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.516: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.519: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:25.534: INFO: Lookups using dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810 failed for: [wheezy_udp@dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_udp@dns-test-service.dns-6099.svc.cluster.local jessie_tcp@dns-test-service.dns-6099.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local] Jun 3 11:29:30.539: INFO: Unable to read wheezy_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.542: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.545: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.549: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.571: INFO: Unable to read jessie_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.574: INFO: Unable to read jessie_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.578: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.581: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:30.601: INFO: Lookups using dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810 failed for: [wheezy_udp@dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_udp@dns-test-service.dns-6099.svc.cluster.local jessie_tcp@dns-test-service.dns-6099.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local] Jun 3 11:29:35.540: INFO: Unable to read wheezy_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.544: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.547: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.551: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.578: INFO: Unable to read jessie_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.581: INFO: Unable to read jessie_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.585: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.589: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:35.612: INFO: Lookups using dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810 failed for: [wheezy_udp@dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_udp@dns-test-service.dns-6099.svc.cluster.local jessie_tcp@dns-test-service.dns-6099.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local] Jun 3 11:29:40.539: INFO: Unable to read wheezy_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.544: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.548: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.552: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.578: INFO: Unable to read jessie_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.582: INFO: Unable to read jessie_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.586: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.589: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:40.611: INFO: Lookups using dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810 failed for: [wheezy_udp@dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_udp@dns-test-service.dns-6099.svc.cluster.local jessie_tcp@dns-test-service.dns-6099.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local] Jun 3 11:29:45.539: INFO: Unable to read wheezy_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.543: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.547: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.551: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.577: INFO: Unable to read jessie_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.581: INFO: Unable to read jessie_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.587: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.591: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:45.610: INFO: Lookups using dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810 failed for: [wheezy_udp@dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_udp@dns-test-service.dns-6099.svc.cluster.local jessie_tcp@dns-test-service.dns-6099.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local] Jun 3 11:29:50.539: INFO: Unable to read wheezy_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.543: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.547: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.551: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.578: INFO: Unable to read jessie_udp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.582: INFO: Unable to read jessie_tcp@dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.586: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.591: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local from pod dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810: the server could not find the requested resource (get pods dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810) Jun 3 11:29:50.612: INFO: Lookups using dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810 failed for: [wheezy_udp@dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@dns-test-service.dns-6099.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_udp@dns-test-service.dns-6099.svc.cluster.local jessie_tcp@dns-test-service.dns-6099.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-6099.svc.cluster.local] Jun 3 11:29:55.611: INFO: DNS probes using dns-6099/dns-test-a68b603f-06fd-40dc-bfbc-11c1792ba810 succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:55.640: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-6099" for this suite. • [SLOW TEST:32.526 seconds] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should provide DNS for services [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for services [Conformance]","total":-1,"completed":15,"skipped":339,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods Extended /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:55.718: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Pods Set QOS Class /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:150 [It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying QOS class is set on the pod [AfterEach] [sig-node] Pods Extended /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:55.756: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-7176" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Pods Extended Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","total":-1,"completed":16,"skipped":379,"failed":0} SSS ------------------------------ [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:55.667: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46 [It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:29:55.704: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-9b375cc9-689a-4a95-a4dc-c57d5b2c8c03" in namespace "security-context-test-6786" to be "Succeeded or Failed" Jun 3 11:29:55.707: INFO: Pod "busybox-privileged-false-9b375cc9-689a-4a95-a4dc-c57d5b2c8c03": Phase="Pending", Reason="", readiness=false. Elapsed: 3.147851ms Jun 3 11:29:57.711: INFO: Pod "busybox-privileged-false-9b375cc9-689a-4a95-a4dc-c57d5b2c8c03": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007544882s Jun 3 11:29:59.715: INFO: Pod "busybox-privileged-false-9b375cc9-689a-4a95-a4dc-c57d5b2c8c03": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011767911s Jun 3 11:29:59.716: INFO: Pod "busybox-privileged-false-9b375cc9-689a-4a95-a4dc-c57d5b2c8c03" satisfied condition "Succeeded or Failed" Jun 3 11:29:59.722: INFO: Got logs for pod "busybox-privileged-false-9b375cc9-689a-4a95-a4dc-c57d5b2c8c03": "ip: RTNETLINK answers: Operation not permitted\n" [AfterEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:29:59.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-6786" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":30,"skipped":395,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:59.773: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0666 on node default medium Jun 3 11:29:59.816: INFO: Waiting up to 5m0s for pod "pod-871ae332-801f-44e5-9218-0165481be31a" in namespace "emptydir-5448" to be "Succeeded or Failed" Jun 3 11:29:59.819: INFO: Pod "pod-871ae332-801f-44e5-9218-0165481be31a": Phase="Pending", Reason="", readiness=false. Elapsed: 3.061054ms Jun 3 11:30:01.823: INFO: Pod "pod-871ae332-801f-44e5-9218-0165481be31a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006915299s Jun 3 11:30:03.828: INFO: Pod "pod-871ae332-801f-44e5-9218-0165481be31a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011488652s STEP: Saw pod success Jun 3 11:30:03.828: INFO: Pod "pod-871ae332-801f-44e5-9218-0165481be31a" satisfied condition "Succeeded or Failed" Jun 3 11:30:03.831: INFO: Trying to get logs from node v1.21-worker pod pod-871ae332-801f-44e5-9218-0165481be31a container test-container: STEP: delete the pod Jun 3 11:30:03.847: INFO: Waiting for pod pod-871ae332-801f-44e5-9218-0165481be31a to disappear Jun 3 11:30:03.852: INFO: Pod pod-871ae332-801f-44e5-9218-0165481be31a no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:03.852: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-5448" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":31,"skipped":416,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:03.875: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should check if kubectl diff finds a difference for Deployments [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create deployment with httpd image Jun 3 11:30:03.907: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-6699 create -f -' Jun 3 11:30:04.321: INFO: stderr: "" Jun 3 11:30:04.321: INFO: stdout: "deployment.apps/httpd-deployment created\n" STEP: verify diff finds difference between live and declared image Jun 3 11:30:04.321: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-6699 diff -f -' Jun 3 11:30:04.637: INFO: rc: 1 Jun 3 11:30:04.637: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-6699 delete -f -' Jun 3 11:30:04.760: INFO: stderr: "" Jun 3 11:30:04.760: INFO: stdout: "deployment.apps \"httpd-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:04.760: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6699" for this suite. • ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl diff should check if kubectl diff finds a difference for Deployments [Conformance]","total":-1,"completed":32,"skipped":422,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:04.775: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should support --unix-socket=/path [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Starting the proxy Jun 3 11:30:04.799: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-4896 proxy --unix-socket=/tmp/kubectl-proxy-unix559945421/test' STEP: retrieving proxy /api/ output [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:04.888: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4896" for this suite. • ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance]","total":-1,"completed":33,"skipped":426,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:17.089: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/init_container.go:162 [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod Jun 3 11:29:17.119: INFO: PodSpec: initContainers in spec.initContainers Jun 3 11:30:07.100: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-50950e48-39f4-44fe-9ca5-be9adb492df6", GenerateName:"", Namespace:"init-container-2955", SelfLink:"", UID:"8ec825ca-da90-488f-8aee-2bc648991c84", ResourceVersion:"3292047", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63758316557, loc:(*time.Location)(0x9dc0820)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"119464363"}, Annotations:map[string]string{"k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.244.1.16\"\n ],\n \"mac\": \"ba:d2:c2:df:b3:43\",\n \"default\": true,\n \"dns\": {}\n}]", "k8s.v1.cni.cncf.io/networks-status":"[{\n \"name\": \"\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.244.1.16\"\n ],\n \"mac\": \"ba:d2:c2:df:b3:43\",\n \"default\": true,\n \"dns\": {}\n}]"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc0020ab2d8), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc0020ab2f0)}, v1.ManagedFieldsEntry{Manager:"multus", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc0020ab308), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc0020ab320)}, v1.ManagedFieldsEntry{Manager:"kubelet", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc0020ab338), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc0020ab350)}}}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-api-access-wfwsm", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(0xc004b1a100), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil), Ephemeral:(*v1.EphemeralVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"k8s.gcr.io/e2e-test-images/busybox:1.29-1", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-wfwsm", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"k8s.gcr.io/e2e-test-images/busybox:1.29-1", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-wfwsm", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.4.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-wfwsm", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc004fdaad8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"v1.21-worker", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc003091c70), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc004fdab90)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc004fdabb0)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc004fdabb8), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc004fdabbc), PreemptionPolicy:(*v1.PreemptionPolicy)(0xc0038f39a0), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil), SetHostnameAsFQDN:(*bool)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316557, loc:(*time.Location)(0x9dc0820)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316557, loc:(*time.Location)(0x9dc0820)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316557, loc:(*time.Location)(0x9dc0820)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316557, loc:(*time.Location)(0x9dc0820)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"172.18.0.4", PodIP:"10.244.1.16", PodIPs:[]v1.PodIP{v1.PodIP{IP:"10.244.1.16"}}, StartTime:(*v1.Time)(0xc0020ab380), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc003091dc0)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc003091e30)}, Ready:false, RestartCount:3, Image:"k8s.gcr.io/e2e-test-images/busybox:1.29-1", ImageID:"k8s.gcr.io/e2e-test-images/busybox@sha256:39e1e963e5310e9c313bad51523be012ede7b35bb9316517d19089a010356592", ContainerID:"containerd://a20d3e0cae45119374546ca91a9c9a9492b1f3536c3396409025f717c37b8e9f", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc004b1a180), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/e2e-test-images/busybox:1.29-1", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc004b1a160), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.4.1", ImageID:"", ContainerID:"", Started:(*bool)(0xc004fdac5f)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}} [AfterEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:07.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-2955" for this suite. • [SLOW TEST:50.021 seconds] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should not start app containers if init containers fail on a RestartAlways pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","total":-1,"completed":10,"skipped":155,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:07.173: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should provide container's memory limit [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:30:07.222: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f4790d71-5298-4f14-bfc9-916d7bf13a13" in namespace "downward-api-4133" to be "Succeeded or Failed" Jun 3 11:30:07.225: INFO: Pod "downwardapi-volume-f4790d71-5298-4f14-bfc9-916d7bf13a13": Phase="Pending", Reason="", readiness=false. Elapsed: 3.104158ms Jun 3 11:30:09.229: INFO: Pod "downwardapi-volume-f4790d71-5298-4f14-bfc9-916d7bf13a13": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007116999s STEP: Saw pod success Jun 3 11:30:09.229: INFO: Pod "downwardapi-volume-f4790d71-5298-4f14-bfc9-916d7bf13a13" satisfied condition "Succeeded or Failed" Jun 3 11:30:09.232: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-f4790d71-5298-4f14-bfc9-916d7bf13a13 container client-container: STEP: delete the pod Jun 3 11:30:09.246: INFO: Waiting for pod downwardapi-volume-f4790d71-5298-4f14-bfc9-916d7bf13a13 to disappear Jun 3 11:30:09.249: INFO: Pod downwardapi-volume-f4790d71-5298-4f14-bfc9-916d7bf13a13 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:09.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-4133" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","total":-1,"completed":11,"skipped":188,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:09.271: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should find a service from listing all namespaces [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: fetching services [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:09.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-9502" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • ------------------------------ {"msg":"PASSED [sig-network] Services should find a service from listing all namespaces [Conformance]","total":-1,"completed":12,"skipped":194,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:09.340: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test override command Jun 3 11:30:09.378: INFO: Waiting up to 5m0s for pod "client-containers-4797f109-c6b1-427e-be8d-04c7945522ad" in namespace "containers-1578" to be "Succeeded or Failed" Jun 3 11:30:09.381: INFO: Pod "client-containers-4797f109-c6b1-427e-be8d-04c7945522ad": Phase="Pending", Reason="", readiness=false. Elapsed: 3.305686ms Jun 3 11:30:11.387: INFO: Pod "client-containers-4797f109-c6b1-427e-be8d-04c7945522ad": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.00861048s STEP: Saw pod success Jun 3 11:30:11.387: INFO: Pod "client-containers-4797f109-c6b1-427e-be8d-04c7945522ad" satisfied condition "Succeeded or Failed" Jun 3 11:30:11.390: INFO: Trying to get logs from node v1.21-worker pod client-containers-4797f109-c6b1-427e-be8d-04c7945522ad container agnhost-container: STEP: delete the pod Jun 3 11:30:11.406: INFO: Waiting for pod client-containers-4797f109-c6b1-427e-be8d-04c7945522ad to disappear Jun 3 11:30:11.411: INFO: Pod client-containers-4797f109-c6b1-427e-be8d-04c7945522ad no longer exists [AfterEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:11.411: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-1578" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":-1,"completed":13,"skipped":203,"failed":0} SS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:11.427: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:30:11.472: INFO: Waiting up to 5m0s for pod "downwardapi-volume-de9c5bc9-946a-485d-8c34-eb63e6da064c" in namespace "downward-api-9790" to be "Succeeded or Failed" Jun 3 11:30:11.475: INFO: Pod "downwardapi-volume-de9c5bc9-946a-485d-8c34-eb63e6da064c": Phase="Pending", Reason="", readiness=false. Elapsed: 3.501414ms Jun 3 11:30:13.480: INFO: Pod "downwardapi-volume-de9c5bc9-946a-485d-8c34-eb63e6da064c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.008589862s STEP: Saw pod success Jun 3 11:30:13.480: INFO: Pod "downwardapi-volume-de9c5bc9-946a-485d-8c34-eb63e6da064c" satisfied condition "Succeeded or Failed" Jun 3 11:30:13.484: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-de9c5bc9-946a-485d-8c34-eb63e6da064c container client-container: STEP: delete the pod Jun 3 11:30:13.499: INFO: Waiting for pod downwardapi-volume-de9c5bc9-946a-485d-8c34-eb63e6da064c to disappear Jun 3 11:30:13.507: INFO: Pod downwardapi-volume-de9c5bc9-946a-485d-8c34-eb63e6da064c no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:13.507: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-9790" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":205,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:13.529: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-volume-map-63de7cbf-1f24-432d-badf-addd6284837c STEP: Creating a pod to test consume configMaps Jun 3 11:30:13.573: INFO: Waiting up to 5m0s for pod "pod-configmaps-63953c6e-302a-46f7-bd72-d8f5ecb3fdf2" in namespace "configmap-9421" to be "Succeeded or Failed" Jun 3 11:30:13.576: INFO: Pod "pod-configmaps-63953c6e-302a-46f7-bd72-d8f5ecb3fdf2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.913351ms Jun 3 11:30:15.581: INFO: Pod "pod-configmaps-63953c6e-302a-46f7-bd72-d8f5ecb3fdf2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.00758892s STEP: Saw pod success Jun 3 11:30:15.581: INFO: Pod "pod-configmaps-63953c6e-302a-46f7-bd72-d8f5ecb3fdf2" satisfied condition "Succeeded or Failed" Jun 3 11:30:15.584: INFO: Trying to get logs from node v1.21-worker pod pod-configmaps-63953c6e-302a-46f7-bd72-d8f5ecb3fdf2 container agnhost-container: STEP: delete the pod Jun 3 11:30:15.600: INFO: Waiting for pod pod-configmaps-63953c6e-302a-46f7-bd72-d8f5ecb3fdf2 to disappear Jun 3 11:30:15.604: INFO: Pod pod-configmaps-63953c6e-302a-46f7-bd72-d8f5ecb3fdf2 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:15.604: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-9421" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":-1,"completed":15,"skipped":210,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:15.645: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the container STEP: wait for the container to reach Failed STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Jun 3 11:30:17.700: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:17.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-2142" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":-1,"completed":16,"skipped":225,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:11.651: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-upd-62820ebe-cab2-4a5b-ad8b-2b91bc56baa5 STEP: Creating the pod Jun 3 11:29:11.800: INFO: The status of Pod pod-configmaps-e86b013d-7c41-4985-ab4f-151b0e48370b is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:13.804: INFO: The status of Pod pod-configmaps-e86b013d-7c41-4985-ab4f-151b0e48370b is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:15.879: INFO: The status of Pod pod-configmaps-e86b013d-7c41-4985-ab4f-151b0e48370b is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:17.804: INFO: The status of Pod pod-configmaps-e86b013d-7c41-4985-ab4f-151b0e48370b is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:29:19.804: INFO: The status of Pod pod-configmaps-e86b013d-7c41-4985-ab4f-151b0e48370b is Running (Ready = true) STEP: Updating configmap configmap-test-upd-62820ebe-cab2-4a5b-ad8b-2b91bc56baa5 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:20.091: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-384" for this suite. • [SLOW TEST:68.451 seconds] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":19,"skipped":419,"failed":0} SSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:17.740: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:30:17.777: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac" in namespace "projected-5023" to be "Succeeded or Failed" Jun 3 11:30:17.780: INFO: Pod "downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac": Phase="Pending", Reason="", readiness=false. Elapsed: 3.177156ms Jun 3 11:30:19.785: INFO: Pod "downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac": Phase="Running", Reason="", readiness=true. Elapsed: 2.007939629s Jun 3 11:30:21.789: INFO: Pod "downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012220261s STEP: Saw pod success Jun 3 11:30:21.789: INFO: Pod "downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac" satisfied condition "Succeeded or Failed" Jun 3 11:30:21.792: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac container client-container: STEP: delete the pod Jun 3 11:30:21.880: INFO: Waiting for pod downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac to disappear Jun 3 11:30:21.883: INFO: Pod downwardapi-volume-f4bc5f9a-ba1a-4b64-8127-5e2abe40a0ac no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:21.883: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5023" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":17,"skipped":234,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] PodTemplates /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:21.934: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename podtemplate STEP: Waiting for a default service account to be provisioned in namespace [It] should run the lifecycle of PodTemplates [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-node] PodTemplates /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:22.001: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "podtemplate-7011" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance]","total":-1,"completed":18,"skipped":251,"failed":0} S ------------------------------ [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:20.130: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for pods for Hostname [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-2520.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-2520.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2520.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-2520.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-2520.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2520.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:30:24.221: INFO: DNS probes using dns-2520/dns-test-03f409c4-c123-4058-a025-0151c97fea71 succeeded STEP: deleting the pod STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:24.236: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-2520" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance]","total":-1,"completed":20,"skipped":432,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:24.258: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1859.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-1859.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1859.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1859.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-1859.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1859.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe /etc/hosts STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:30:26.342: INFO: DNS probes using dns-1859/dns-test-016011bd-4892-428a-9c48-761e269324b3 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:26.351: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-1859" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]","total":-1,"completed":21,"skipped":438,"failed":0} SS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:26.367: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should be able to change the type from ExternalName to NodePort [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a service externalname-service with the type=ExternalName in namespace services-9030 STEP: changing the ExternalName service to type=NodePort STEP: creating replication controller externalname-service in namespace services-9030 I0603 11:30:26.428291 34 runners.go:190] Created replication controller with name: externalname-service, namespace: services-9030, replica count: 2 I0603 11:30:29.478984 34 runners.go:190] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:30:29.479: INFO: Creating new exec pod Jun 3 11:30:32.496: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-9030 exec execpods5zrs -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' Jun 3 11:30:32.751: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" Jun 3 11:30:32.751: INFO: stdout: "externalname-service-xrt77" Jun 3 11:30:32.751: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-9030 exec execpods5zrs -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.230.179 80' Jun 3 11:30:32.988: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.230.179 80\nConnection to 10.96.230.179 80 port [tcp/http] succeeded!\n" Jun 3 11:30:32.988: INFO: stdout: "externalname-service-x2qt6" Jun 3 11:30:32.988: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-9030 exec execpods5zrs -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.4 31520' Jun 3 11:30:33.230: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.4 31520\nConnection to 172.18.0.4 31520 port [tcp/*] succeeded!\n" Jun 3 11:30:33.230: INFO: stdout: "externalname-service-xrt77" Jun 3 11:30:33.230: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-9030 exec execpods5zrs -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.2 31520' Jun 3 11:30:33.443: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.2 31520\nConnection to 172.18.0.2 31520 port [tcp/*] succeeded!\n" Jun 3 11:30:33.444: INFO: stdout: "externalname-service-xrt77" Jun 3 11:30:33.444: INFO: Cleaning up the ExternalName to NodePort test service [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:33.464: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-9030" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:7.104 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should be able to change the type from ExternalName to NodePort [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","total":-1,"completed":22,"skipped":440,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-network] Ingress API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:33.486: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename ingress STEP: Waiting for a default service account to be provisioned in namespace [It] should support creating Ingress API operations [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: getting /apis STEP: getting /apis/networking.k8s.io STEP: getting /apis/networking.k8s.iov1 STEP: creating STEP: getting STEP: listing STEP: watching Jun 3 11:30:33.547: INFO: starting watch STEP: cluster-wide listing STEP: cluster-wide watching Jun 3 11:30:33.551: INFO: starting watch STEP: patching STEP: updating Jun 3 11:30:33.564: INFO: waiting for watch events with expected annotations Jun 3 11:30:33.564: INFO: missing expected annotations, waiting: map[string]string(nil) Jun 3 11:30:33.564: INFO: saw patched and updated annotations STEP: patching /status STEP: updating /status STEP: get /status STEP: deleting STEP: deleting a collection [AfterEach] [sig-network] Ingress API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:33.620: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "ingress-3819" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] Ingress API should support creating Ingress API operations [Conformance]","total":-1,"completed":23,"skipped":448,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:55.193: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for ExternalName services [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a test externalName service STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5330.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5330.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local; sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:29:59.263: INFO: DNS probes using dns-test-b895b78f-b047-4b63-99fc-65ed4708aab5 succeeded STEP: deleting the pod STEP: changing the externalName to bar.example.com STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5330.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5330.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local; sleep 1; done STEP: creating a second pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:30:03.307: INFO: File wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:03.313: INFO: File jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:03.313: INFO: Lookups using dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 failed for: [wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local] Jun 3 11:30:08.319: INFO: File wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:08.323: INFO: File jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:08.323: INFO: Lookups using dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 failed for: [wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local] Jun 3 11:30:13.320: INFO: File wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:13.324: INFO: File jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:13.324: INFO: Lookups using dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 failed for: [wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local] Jun 3 11:30:18.380: INFO: File wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:18.482: INFO: File jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:18.482: INFO: Lookups using dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 failed for: [wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local] Jun 3 11:30:23.319: INFO: File wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:23.323: INFO: File jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:23.323: INFO: Lookups using dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 failed for: [wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local] Jun 3 11:30:28.319: INFO: File wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local from pod dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 contains 'foo.example.com. ' instead of 'bar.example.com.' Jun 3 11:30:28.323: INFO: Lookups using dns-5330/dns-test-d8363004-9698-451e-a052-861b23c26a32 failed for: [wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local] Jun 3 11:30:33.323: INFO: DNS probes using dns-test-d8363004-9698-451e-a052-861b23c26a32 succeeded STEP: deleting the pod STEP: changing the service to type=ClusterIP STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5330.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-5330.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5330.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-5330.svc.cluster.local; sleep 1; done STEP: creating a third pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:30:35.378: INFO: DNS probes using dns-test-abe96144-7b8a-4f71-8dfd-497fda4f5d40 succeeded STEP: deleting the pod STEP: deleting the test externalName service [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:35.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-5330" for this suite. • [SLOW TEST:40.212 seconds] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should provide DNS for ExternalName services [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","total":-1,"completed":13,"skipped":265,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:33.656: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward api env vars Jun 3 11:30:33.690: INFO: Waiting up to 5m0s for pod "downward-api-f9013955-5051-40af-a26c-f25e56c25189" in namespace "downward-api-1794" to be "Succeeded or Failed" Jun 3 11:30:33.694: INFO: Pod "downward-api-f9013955-5051-40af-a26c-f25e56c25189": Phase="Pending", Reason="", readiness=false. Elapsed: 3.954104ms Jun 3 11:30:35.698: INFO: Pod "downward-api-f9013955-5051-40af-a26c-f25e56c25189": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007421993s STEP: Saw pod success Jun 3 11:30:35.698: INFO: Pod "downward-api-f9013955-5051-40af-a26c-f25e56c25189" satisfied condition "Succeeded or Failed" Jun 3 11:30:35.701: INFO: Trying to get logs from node v1.21-worker2 pod downward-api-f9013955-5051-40af-a26c-f25e56c25189 container dapi-container: STEP: delete the pod Jun 3 11:30:35.715: INFO: Waiting for pod downward-api-f9013955-5051-40af-a26c-f25e56c25189 to disappear Jun 3 11:30:35.718: INFO: Pod downward-api-f9013955-5051-40af-a26c-f25e56c25189 no longer exists [AfterEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:35.718: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1794" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","total":-1,"completed":24,"skipped":460,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:35.452: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Jun 3 11:30:37.501: INFO: Expected: &{} to match Container's Termination Message: -- STEP: delete the container [AfterEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:37.514: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-9225" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":292,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:35.743: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a replication controller. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ReplicationController STEP: Ensuring resource quota status captures replication controller creation STEP: Deleting a ReplicationController STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:47.296: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-618" for this suite. • [SLOW TEST:11.564 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a replication controller. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","total":-1,"completed":25,"skipped":469,"failed":0} SSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:47.337: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0777 on tmpfs Jun 3 11:30:47.380: INFO: Waiting up to 5m0s for pod "pod-effc431f-5535-48a2-9f94-9fe3a7709968" in namespace "emptydir-2274" to be "Succeeded or Failed" Jun 3 11:30:47.383: INFO: Pod "pod-effc431f-5535-48a2-9f94-9fe3a7709968": Phase="Pending", Reason="", readiness=false. Elapsed: 2.573474ms Jun 3 11:30:49.387: INFO: Pod "pod-effc431f-5535-48a2-9f94-9fe3a7709968": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007043449s STEP: Saw pod success Jun 3 11:30:49.388: INFO: Pod "pod-effc431f-5535-48a2-9f94-9fe3a7709968" satisfied condition "Succeeded or Failed" Jun 3 11:30:49.390: INFO: Trying to get logs from node v1.21-worker2 pod pod-effc431f-5535-48a2-9f94-9fe3a7709968 container test-container: STEP: delete the pod Jun 3 11:30:49.404: INFO: Waiting for pod pod-effc431f-5535-48a2-9f94-9fe3a7709968 to disappear Jun 3 11:30:49.407: INFO: Pod pod-effc431f-5535-48a2-9f94-9fe3a7709968 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:49.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2274" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":26,"skipped":483,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:49.483: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating projection with secret that has name projected-secret-test-a1667f6a-7679-41d6-9da1-96bc101d7e6b STEP: Creating a pod to test consume secrets Jun 3 11:30:49.529: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-475afa45-27b6-4463-8019-41604e24cd91" in namespace "projected-4433" to be "Succeeded or Failed" Jun 3 11:30:49.532: INFO: Pod "pod-projected-secrets-475afa45-27b6-4463-8019-41604e24cd91": Phase="Pending", Reason="", readiness=false. Elapsed: 3.383004ms Jun 3 11:30:51.537: INFO: Pod "pod-projected-secrets-475afa45-27b6-4463-8019-41604e24cd91": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.008438576s STEP: Saw pod success Jun 3 11:30:51.537: INFO: Pod "pod-projected-secrets-475afa45-27b6-4463-8019-41604e24cd91" satisfied condition "Succeeded or Failed" Jun 3 11:30:51.541: INFO: Trying to get logs from node v1.21-worker2 pod pod-projected-secrets-475afa45-27b6-4463-8019-41604e24cd91 container projected-secret-volume-test: STEP: delete the pod Jun 3 11:30:51.556: INFO: Waiting for pod pod-projected-secrets-475afa45-27b6-4463-8019-41604e24cd91 to disappear Jun 3 11:30:51.559: INFO: Pod pod-projected-secrets-475afa45-27b6-4463-8019-41604e24cd91 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:51.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4433" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","total":-1,"completed":27,"skipped":520,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/sysctl.go:35 [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:51.608: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sysctl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/sysctl.go:64 [It] should support unsafe sysctls which are actually allowed [MinimumKubeletVersion:1.21] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod with the kernel.shm_rmid_forced sysctl STEP: Watching for error events or started pod STEP: Waiting for pod completion STEP: Checking that the pod succeeded STEP: Getting logs from the pod STEP: Checking that the sysctl is actually updated [AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:53.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sysctl-9884" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] should support unsafe sysctls which are actually allowed [MinimumKubeletVersion:1.21] [Conformance]","total":-1,"completed":28,"skipped":539,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:55.774: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a watch on configmaps with label A STEP: creating a watch on configmaps with label B STEP: creating a watch on configmaps with label A or B STEP: creating a configmap with label A and ensuring the correct watchers observe the notification Jun 3 11:29:55.808: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3291795 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:29:55 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:29:55.809: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3291795 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:29:55 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying configmap A and ensuring the correct watchers observe the notification Jun 3 11:30:05.818: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3292035 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:05 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:30:05.818: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3292035 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:05 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying configmap A again and ensuring the correct watchers observe the notification Jun 3 11:30:15.827: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3292225 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:05 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:30:15.827: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3292225 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:05 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: deleting configmap A and ensuring the correct watchers observe the notification Jun 3 11:30:25.834: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3292451 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:05 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:30:25.834: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3664 43c36b3a-69eb-4be1-ac25-ab6b3033f85b 3292451 0 2021-06-03 11:29:55 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:05 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: creating a configmap with label B and ensuring the correct watchers observe the notification Jun 3 11:30:35.842: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3664 645e8698-f812-4576-93e3-7ef38c4711b5 3292740 0 2021-06-03 11:30:35 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:35 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:30:35.843: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3664 645e8698-f812-4576-93e3-7ef38c4711b5 3292740 0 2021-06-03 11:30:35 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:35 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} STEP: deleting configmap B and ensuring the correct watchers observe the notification Jun 3 11:30:45.881: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3664 645e8698-f812-4576-93e3-7ef38c4711b5 3292993 0 2021-06-03 11:30:35 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:35 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:30:45.881: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3664 645e8698-f812-4576-93e3-7ef38c4711b5 3292993 0 2021-06-03 11:30:35 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2021-06-03 11:30:35 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:55.883: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-3664" for this suite. • [SLOW TEST:60.121 seconds] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":-1,"completed":17,"skipped":382,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:53.715: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:30:54.157: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:30:57.379: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API STEP: Creating a dummy validating-webhook-configuration object STEP: Deleting the validating-webhook-configuration, which should be possible to remove STEP: Creating a dummy mutating-webhook-configuration object STEP: Deleting the mutating-webhook-configuration, which should be possible to remove [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:30:57.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2469" for this suite. STEP: Destroying namespace "webhook-2469-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","total":-1,"completed":29,"skipped":544,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:57.762: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:30:59.784: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:31:04.280: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316659, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:31:06.290: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316659, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:31:08.286: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316660, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316659, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:31:11.294: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny pod and configmap creation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering the webhook via the AdmissionRegistration API STEP: create a pod that should be denied by the webhook STEP: create a pod that causes the webhook to hang STEP: create a configmap that should be denied by the webhook STEP: create a configmap that should be admitted by the webhook STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook STEP: create a namespace that bypass the webhook STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:21.428: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-5336" for this suite. STEP: Destroying namespace "webhook-5336-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:23.700 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny pod and configmap creation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","total":-1,"completed":30,"skipped":548,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:40.854: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan pods created by rc if delete options say so [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods STEP: Gathering metrics W0603 11:30:20.991803 32 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Jun 3 11:31:23.011: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. Jun 3 11:31:23.011: INFO: Deleting pod "simpletest.rc-2n84r" in namespace "gc-5221" Jun 3 11:31:23.021: INFO: Deleting pod "simpletest.rc-72bpp" in namespace "gc-5221" Jun 3 11:31:23.030: INFO: Deleting pod "simpletest.rc-8vkx9" in namespace "gc-5221" Jun 3 11:31:23.038: INFO: Deleting pod "simpletest.rc-bdqx2" in namespace "gc-5221" Jun 3 11:31:23.045: INFO: Deleting pod "simpletest.rc-gd8gk" in namespace "gc-5221" Jun 3 11:31:23.052: INFO: Deleting pod "simpletest.rc-lhp9c" in namespace "gc-5221" Jun 3 11:31:23.060: INFO: Deleting pod "simpletest.rc-m42bx" in namespace "gc-5221" Jun 3 11:31:23.066: INFO: Deleting pod "simpletest.rc-njnf7" in namespace "gc-5221" Jun 3 11:31:23.073: INFO: Deleting pod "simpletest.rc-xhdhc" in namespace "gc-5221" Jun 3 11:31:23.086: INFO: Deleting pod "simpletest.rc-zzscw" in namespace "gc-5221" [AfterEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:23.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-5221" for this suite. • [SLOW TEST:102.344 seconds] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan pods created by rc if delete options say so [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","total":-1,"completed":17,"skipped":400,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:23.314: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should delete a collection of pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Create set of pods Jun 3 11:31:23.360: INFO: created test-pod-1 Jun 3 11:31:23.364: INFO: created test-pod-2 Jun 3 11:31:23.367: INFO: created test-pod-3 STEP: waiting for all 3 pods to be located STEP: waiting for all pods to be deleted [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:23.390: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8092" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Pods should delete a collection of pods [Conformance]","total":-1,"completed":18,"skipped":465,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:21.482: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name projected-secret-test-591a2e86-a90d-46c2-97fc-346c3845cc2d STEP: Creating a pod to test consume secrets Jun 3 11:31:21.523: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-24d04911-1e96-4157-b8db-2d86b6e9dbf0" in namespace "projected-1500" to be "Succeeded or Failed" Jun 3 11:31:21.526: INFO: Pod "pod-projected-secrets-24d04911-1e96-4157-b8db-2d86b6e9dbf0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.979725ms Jun 3 11:31:23.530: INFO: Pod "pod-projected-secrets-24d04911-1e96-4157-b8db-2d86b6e9dbf0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007291379s STEP: Saw pod success Jun 3 11:31:23.530: INFO: Pod "pod-projected-secrets-24d04911-1e96-4157-b8db-2d86b6e9dbf0" satisfied condition "Succeeded or Failed" Jun 3 11:31:23.533: INFO: Trying to get logs from node v1.21-worker pod pod-projected-secrets-24d04911-1e96-4157-b8db-2d86b6e9dbf0 container secret-volume-test: STEP: delete the pod Jun 3 11:31:23.545: INFO: Waiting for pod pod-projected-secrets-24d04911-1e96-4157-b8db-2d86b6e9dbf0 to disappear Jun 3 11:31:23.547: INFO: Pod pod-projected-secrets-24d04911-1e96-4157-b8db-2d86b6e9dbf0 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:23.547: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1500" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":-1,"completed":31,"skipped":557,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:42.170: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105 STEP: Creating service test in namespace statefulset-3275 [It] should perform rolling updates and roll backs of template modifications [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a new StatefulSet Jun 3 11:28:42.209: INFO: Found 0 stateful pods, waiting for 3 Jun 3 11:28:52.214: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:28:52.214: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:28:52.214: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Jun 3 11:29:02.214: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:29:02.214: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:29:02.214: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:29:02.223: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-3275 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:29:02.471: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:29:02.471: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:29:02.471: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' STEP: Updating StatefulSet template: update image from k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 to k8s.gcr.io/e2e-test-images/httpd:2.4.39-1 Jun 3 11:29:12.508: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Updating Pods in reverse ordinal order Jun 3 11:29:22.526: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-3275 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:29:22.766: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 3 11:29:22.766: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:29:22.766: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:29:32.790: INFO: Waiting for StatefulSet statefulset-3275/ss2 to complete update Jun 3 11:29:32.790: INFO: Waiting for Pod statefulset-3275/ss2-0 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 Jun 3 11:29:32.790: INFO: Waiting for Pod statefulset-3275/ss2-1 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 Jun 3 11:29:32.790: INFO: Waiting for Pod statefulset-3275/ss2-2 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 Jun 3 11:29:42.799: INFO: Waiting for StatefulSet statefulset-3275/ss2 to complete update Jun 3 11:29:42.799: INFO: Waiting for Pod statefulset-3275/ss2-0 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 Jun 3 11:29:42.799: INFO: Waiting for Pod statefulset-3275/ss2-1 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 Jun 3 11:29:52.799: INFO: Waiting for StatefulSet statefulset-3275/ss2 to complete update Jun 3 11:29:52.800: INFO: Waiting for Pod statefulset-3275/ss2-0 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 STEP: Rolling back to a previous revision Jun 3 11:30:02.798: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-3275 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:30:03.050: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:30:03.050: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:30:03.050: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:30:13.086: INFO: Updating stateful set ss2 STEP: Rolling back update in reverse ordinal order Jun 3 11:30:23.106: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-3275 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:30:23.325: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 3 11:30:23.325: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:30:23.325: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:30:43.584: INFO: Waiting for StatefulSet statefulset-3275/ss2 to complete update Jun 3 11:30:43.584: INFO: Waiting for Pod statefulset-3275/ss2-0 to have revision ss2-677d6db895 update revision ss2-5bbbc9fc94 [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116 Jun 3 11:30:53.685: INFO: Deleting all statefulset in ns statefulset-3275 Jun 3 11:30:53.689: INFO: Scaling statefulset ss2 to 0 Jun 3 11:31:23.705: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:31:23.709: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:23.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-3275" for this suite. • [SLOW TEST:161.646 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95 should perform rolling updates and roll backs of template modifications [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","total":-1,"completed":12,"skipped":163,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:23.409: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should add annotations for pods in rc [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating Agnhost RC Jun 3 11:31:23.499: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9653 create -f -' Jun 3 11:31:23.914: INFO: stderr: "" Jun 3 11:31:23.914: INFO: stdout: "replicationcontroller/agnhost-primary created\n" STEP: Waiting for Agnhost primary to start. Jun 3 11:31:24.919: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:31:24.919: INFO: Found 0 / 1 Jun 3 11:31:25.918: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:31:25.919: INFO: Found 1 / 1 Jun 3 11:31:25.919: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 STEP: patching all pods Jun 3 11:31:25.922: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:31:25.922: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Jun 3 11:31:25.922: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9653 patch pod agnhost-primary-m6x4w -p {"metadata":{"annotations":{"x":"y"}}}' Jun 3 11:31:26.052: INFO: stderr: "" Jun 3 11:31:26.052: INFO: stdout: "pod/agnhost-primary-m6x4w patched\n" STEP: checking annotations Jun 3 11:31:26.056: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:31:26.056: INFO: ForEach: Found 1 pods from the filter. Now looping through them. [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:26.056: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9653" for this suite. • ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance]","total":-1,"completed":19,"skipped":472,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:23.582: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:31:23.620: INFO: Waiting up to 5m0s for pod "downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92" in namespace "projected-8558" to be "Succeeded or Failed" Jun 3 11:31:23.622: INFO: Pod "downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92": Phase="Pending", Reason="", readiness=false. Elapsed: 2.477449ms Jun 3 11:31:25.626: INFO: Pod "downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006509081s Jun 3 11:31:27.631: INFO: Pod "downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01130183s STEP: Saw pod success Jun 3 11:31:27.631: INFO: Pod "downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92" satisfied condition "Succeeded or Failed" Jun 3 11:31:27.634: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92 container client-container: STEP: delete the pod Jun 3 11:31:27.650: INFO: Waiting for pod downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92 to disappear Jun 3 11:31:27.653: INFO: Pod downwardapi-volume-478fa7ff-23ce-4416-971e-13c1102a2f92 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:27.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8558" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":-1,"completed":32,"skipped":574,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:23.829: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Jun 3 11:31:27.888: INFO: Expected: &{OK} to match Container's Termination Message: OK -- STEP: delete the container [AfterEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:27.899: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-2060" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":-1,"completed":13,"skipped":171,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Discovery /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:27.677: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename discovery STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Discovery /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/discovery.go:39 STEP: Setting up server cert [It] should validate PreferredVersion for each APIGroup [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:31:28.213: INFO: Checking APIGroup: apiregistration.k8s.io Jun 3 11:31:28.214: INFO: PreferredVersion.GroupVersion: apiregistration.k8s.io/v1 Jun 3 11:31:28.214: INFO: Versions found [{apiregistration.k8s.io/v1 v1} {apiregistration.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.214: INFO: apiregistration.k8s.io/v1 matches apiregistration.k8s.io/v1 Jun 3 11:31:28.214: INFO: Checking APIGroup: apps Jun 3 11:31:28.215: INFO: PreferredVersion.GroupVersion: apps/v1 Jun 3 11:31:28.215: INFO: Versions found [{apps/v1 v1}] Jun 3 11:31:28.215: INFO: apps/v1 matches apps/v1 Jun 3 11:31:28.215: INFO: Checking APIGroup: events.k8s.io Jun 3 11:31:28.216: INFO: PreferredVersion.GroupVersion: events.k8s.io/v1 Jun 3 11:31:28.216: INFO: Versions found [{events.k8s.io/v1 v1} {events.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.216: INFO: events.k8s.io/v1 matches events.k8s.io/v1 Jun 3 11:31:28.216: INFO: Checking APIGroup: authentication.k8s.io Jun 3 11:31:28.218: INFO: PreferredVersion.GroupVersion: authentication.k8s.io/v1 Jun 3 11:31:28.218: INFO: Versions found [{authentication.k8s.io/v1 v1} {authentication.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.218: INFO: authentication.k8s.io/v1 matches authentication.k8s.io/v1 Jun 3 11:31:28.218: INFO: Checking APIGroup: authorization.k8s.io Jun 3 11:31:28.219: INFO: PreferredVersion.GroupVersion: authorization.k8s.io/v1 Jun 3 11:31:28.219: INFO: Versions found [{authorization.k8s.io/v1 v1} {authorization.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.219: INFO: authorization.k8s.io/v1 matches authorization.k8s.io/v1 Jun 3 11:31:28.219: INFO: Checking APIGroup: autoscaling Jun 3 11:31:28.220: INFO: PreferredVersion.GroupVersion: autoscaling/v1 Jun 3 11:31:28.220: INFO: Versions found [{autoscaling/v1 v1} {autoscaling/v2beta1 v2beta1} {autoscaling/v2beta2 v2beta2}] Jun 3 11:31:28.220: INFO: autoscaling/v1 matches autoscaling/v1 Jun 3 11:31:28.220: INFO: Checking APIGroup: batch Jun 3 11:31:28.220: INFO: PreferredVersion.GroupVersion: batch/v1 Jun 3 11:31:28.220: INFO: Versions found [{batch/v1 v1} {batch/v1beta1 v1beta1}] Jun 3 11:31:28.221: INFO: batch/v1 matches batch/v1 Jun 3 11:31:28.221: INFO: Checking APIGroup: certificates.k8s.io Jun 3 11:31:28.221: INFO: PreferredVersion.GroupVersion: certificates.k8s.io/v1 Jun 3 11:31:28.221: INFO: Versions found [{certificates.k8s.io/v1 v1} {certificates.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.221: INFO: certificates.k8s.io/v1 matches certificates.k8s.io/v1 Jun 3 11:31:28.221: INFO: Checking APIGroup: networking.k8s.io Jun 3 11:31:28.222: INFO: PreferredVersion.GroupVersion: networking.k8s.io/v1 Jun 3 11:31:28.222: INFO: Versions found [{networking.k8s.io/v1 v1} {networking.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.223: INFO: networking.k8s.io/v1 matches networking.k8s.io/v1 Jun 3 11:31:28.223: INFO: Checking APIGroup: extensions Jun 3 11:31:28.224: INFO: PreferredVersion.GroupVersion: extensions/v1beta1 Jun 3 11:31:28.224: INFO: Versions found [{extensions/v1beta1 v1beta1}] Jun 3 11:31:28.224: INFO: extensions/v1beta1 matches extensions/v1beta1 Jun 3 11:31:28.224: INFO: Checking APIGroup: policy Jun 3 11:31:28.225: INFO: PreferredVersion.GroupVersion: policy/v1 Jun 3 11:31:28.225: INFO: Versions found [{policy/v1 v1} {policy/v1beta1 v1beta1}] Jun 3 11:31:28.225: INFO: policy/v1 matches policy/v1 Jun 3 11:31:28.225: INFO: Checking APIGroup: rbac.authorization.k8s.io Jun 3 11:31:28.226: INFO: PreferredVersion.GroupVersion: rbac.authorization.k8s.io/v1 Jun 3 11:31:28.226: INFO: Versions found [{rbac.authorization.k8s.io/v1 v1} {rbac.authorization.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.226: INFO: rbac.authorization.k8s.io/v1 matches rbac.authorization.k8s.io/v1 Jun 3 11:31:28.226: INFO: Checking APIGroup: storage.k8s.io Jun 3 11:31:28.227: INFO: PreferredVersion.GroupVersion: storage.k8s.io/v1 Jun 3 11:31:28.227: INFO: Versions found [{storage.k8s.io/v1 v1} {storage.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.227: INFO: storage.k8s.io/v1 matches storage.k8s.io/v1 Jun 3 11:31:28.227: INFO: Checking APIGroup: admissionregistration.k8s.io Jun 3 11:31:28.228: INFO: PreferredVersion.GroupVersion: admissionregistration.k8s.io/v1 Jun 3 11:31:28.228: INFO: Versions found [{admissionregistration.k8s.io/v1 v1} {admissionregistration.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.228: INFO: admissionregistration.k8s.io/v1 matches admissionregistration.k8s.io/v1 Jun 3 11:31:28.228: INFO: Checking APIGroup: apiextensions.k8s.io Jun 3 11:31:28.229: INFO: PreferredVersion.GroupVersion: apiextensions.k8s.io/v1 Jun 3 11:31:28.229: INFO: Versions found [{apiextensions.k8s.io/v1 v1} {apiextensions.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.229: INFO: apiextensions.k8s.io/v1 matches apiextensions.k8s.io/v1 Jun 3 11:31:28.229: INFO: Checking APIGroup: scheduling.k8s.io Jun 3 11:31:28.230: INFO: PreferredVersion.GroupVersion: scheduling.k8s.io/v1 Jun 3 11:31:28.230: INFO: Versions found [{scheduling.k8s.io/v1 v1} {scheduling.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.230: INFO: scheduling.k8s.io/v1 matches scheduling.k8s.io/v1 Jun 3 11:31:28.230: INFO: Checking APIGroup: coordination.k8s.io Jun 3 11:31:28.231: INFO: PreferredVersion.GroupVersion: coordination.k8s.io/v1 Jun 3 11:31:28.232: INFO: Versions found [{coordination.k8s.io/v1 v1} {coordination.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.232: INFO: coordination.k8s.io/v1 matches coordination.k8s.io/v1 Jun 3 11:31:28.232: INFO: Checking APIGroup: node.k8s.io Jun 3 11:31:28.233: INFO: PreferredVersion.GroupVersion: node.k8s.io/v1 Jun 3 11:31:28.233: INFO: Versions found [{node.k8s.io/v1 v1} {node.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.233: INFO: node.k8s.io/v1 matches node.k8s.io/v1 Jun 3 11:31:28.233: INFO: Checking APIGroup: discovery.k8s.io Jun 3 11:31:28.235: INFO: PreferredVersion.GroupVersion: discovery.k8s.io/v1 Jun 3 11:31:28.235: INFO: Versions found [{discovery.k8s.io/v1 v1} {discovery.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.235: INFO: discovery.k8s.io/v1 matches discovery.k8s.io/v1 Jun 3 11:31:28.235: INFO: Checking APIGroup: flowcontrol.apiserver.k8s.io Jun 3 11:31:28.236: INFO: PreferredVersion.GroupVersion: flowcontrol.apiserver.k8s.io/v1beta1 Jun 3 11:31:28.236: INFO: Versions found [{flowcontrol.apiserver.k8s.io/v1beta1 v1beta1}] Jun 3 11:31:28.236: INFO: flowcontrol.apiserver.k8s.io/v1beta1 matches flowcontrol.apiserver.k8s.io/v1beta1 Jun 3 11:31:28.236: INFO: Checking APIGroup: k8s.cni.cncf.io Jun 3 11:31:28.237: INFO: PreferredVersion.GroupVersion: k8s.cni.cncf.io/v1 Jun 3 11:31:28.237: INFO: Versions found [{k8s.cni.cncf.io/v1 v1}] Jun 3 11:31:28.237: INFO: k8s.cni.cncf.io/v1 matches k8s.cni.cncf.io/v1 Jun 3 11:31:28.237: INFO: Checking APIGroup: projectcontour.io Jun 3 11:31:28.238: INFO: PreferredVersion.GroupVersion: projectcontour.io/v1 Jun 3 11:31:28.238: INFO: Versions found [{projectcontour.io/v1 v1} {projectcontour.io/v1alpha1 v1alpha1}] Jun 3 11:31:28.238: INFO: projectcontour.io/v1 matches projectcontour.io/v1 Jun 3 11:31:28.238: INFO: Checking APIGroup: litmuschaos.io Jun 3 11:31:28.239: INFO: PreferredVersion.GroupVersion: litmuschaos.io/v1alpha1 Jun 3 11:31:28.239: INFO: Versions found [{litmuschaos.io/v1alpha1 v1alpha1}] Jun 3 11:31:28.239: INFO: litmuschaos.io/v1alpha1 matches litmuschaos.io/v1alpha1 Jun 3 11:31:28.239: INFO: Checking APIGroup: pingcap.com Jun 3 11:31:28.241: INFO: PreferredVersion.GroupVersion: pingcap.com/v1alpha1 Jun 3 11:31:28.241: INFO: Versions found [{pingcap.com/v1alpha1 v1alpha1}] Jun 3 11:31:28.241: INFO: pingcap.com/v1alpha1 matches pingcap.com/v1alpha1 [AfterEach] [sig-api-machinery] Discovery /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:28.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "discovery-8426" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] Discovery should validate PreferredVersion for each APIGroup [Conformance]","total":-1,"completed":33,"skipped":580,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:27.937: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name projected-configmap-test-volume-19d4f48b-22b0-48ed-82e2-a707595653ec STEP: Creating a pod to test consume configMaps Jun 3 11:31:27.973: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8" in namespace "projected-3750" to be "Succeeded or Failed" Jun 3 11:31:27.976: INFO: Pod "pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.7708ms Jun 3 11:31:29.983: INFO: Pod "pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00908359s Jun 3 11:31:31.987: INFO: Pod "pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01382173s STEP: Saw pod success Jun 3 11:31:31.987: INFO: Pod "pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8" satisfied condition "Succeeded or Failed" Jun 3 11:31:31.991: INFO: Trying to get logs from node v1.21-worker pod pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8 container agnhost-container: STEP: delete the pod Jun 3 11:31:32.006: INFO: Waiting for pod pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8 to disappear Jun 3 11:31:32.009: INFO: Pod pod-projected-configmaps-b7841477-c8bc-4e27-a0b4-3e59d575bad8 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:32.009: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3750" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":188,"failed":0} SSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:32.047: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] should include custom resource definition resources in discovery documents [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: fetching the /apis discovery document STEP: finding the apiextensions.k8s.io API group in the /apis discovery document STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document STEP: fetching the /apis/apiextensions.k8s.io discovery document STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:32.081: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-264" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","total":-1,"completed":15,"skipped":199,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:26.164: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod UID as env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward api env vars Jun 3 11:31:26.206: INFO: Waiting up to 5m0s for pod "downward-api-986c44ff-444a-467d-a1b2-e0af836fb811" in namespace "downward-api-5792" to be "Succeeded or Failed" Jun 3 11:31:26.209: INFO: Pod "downward-api-986c44ff-444a-467d-a1b2-e0af836fb811": Phase="Pending", Reason="", readiness=false. Elapsed: 2.878845ms Jun 3 11:31:28.213: INFO: Pod "downward-api-986c44ff-444a-467d-a1b2-e0af836fb811": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007277723s Jun 3 11:31:30.218: INFO: Pod "downward-api-986c44ff-444a-467d-a1b2-e0af836fb811": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011618968s Jun 3 11:31:32.221: INFO: Pod "downward-api-986c44ff-444a-467d-a1b2-e0af836fb811": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015144603s STEP: Saw pod success Jun 3 11:31:32.221: INFO: Pod "downward-api-986c44ff-444a-467d-a1b2-e0af836fb811" satisfied condition "Succeeded or Failed" Jun 3 11:31:32.225: INFO: Trying to get logs from node v1.21-worker pod downward-api-986c44ff-444a-467d-a1b2-e0af836fb811 container dapi-container: STEP: delete the pod Jun 3 11:31:32.238: INFO: Waiting for pod downward-api-986c44ff-444a-467d-a1b2-e0af836fb811 to disappear Jun 3 11:31:32.240: INFO: Pod downward-api-986c44ff-444a-467d-a1b2-e0af836fb811 no longer exists [AfterEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:32.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5792" for this suite. • [SLOW TEST:6.086 seconds] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should provide pod UID as env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":-1,"completed":20,"skipped":521,"failed":0} SS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:22.013: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name cm-test-opt-del-1a3789ca-e4a6-401d-8828-4172259c18c3 STEP: Creating configMap with name cm-test-opt-upd-1b925606-658a-47e7-b882-079c13ac9327 STEP: Creating the pod Jun 3 11:30:22.066: INFO: The status of Pod pod-projected-configmaps-fc112484-5445-4529-a210-4fff93f7a885 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:30:24.071: INFO: The status of Pod pod-projected-configmaps-fc112484-5445-4529-a210-4fff93f7a885 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:30:26.071: INFO: The status of Pod pod-projected-configmaps-fc112484-5445-4529-a210-4fff93f7a885 is Running (Ready = true) STEP: Deleting configmap cm-test-opt-del-1a3789ca-e4a6-401d-8828-4172259c18c3 STEP: Updating configmap cm-test-opt-upd-1b925606-658a-47e7-b882-079c13ac9327 STEP: Creating configMap with name cm-test-opt-create-b038e12c-dd39-490f-ba35-964beb7464b2 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:34.521: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5958" for this suite. • [SLOW TEST:72.516 seconds] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":19,"skipped":252,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:37.542: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service in namespace services-6086 Jun 3 11:30:37.588: INFO: The status of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:30:39.593: INFO: The status of Pod kube-proxy-mode-detector is Running (Ready = true) Jun 3 11:30:39.596: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6086 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' Jun 3 11:30:39.866: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" Jun 3 11:30:39.866: INFO: stdout: "iptables" Jun 3 11:30:39.866: INFO: proxyMode: iptables Jun 3 11:30:39.883: INFO: Waiting for pod kube-proxy-mode-detector to disappear Jun 3 11:30:39.886: INFO: Pod kube-proxy-mode-detector no longer exists STEP: creating service affinity-clusterip-timeout in namespace services-6086 STEP: creating replication controller affinity-clusterip-timeout in namespace services-6086 I0603 11:30:39.898439 17 runners.go:190] Created replication controller with name: affinity-clusterip-timeout, namespace: services-6086, replica count: 3 I0603 11:30:42.949652 17 runners.go:190] affinity-clusterip-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:30:43.280: INFO: Creating new exec pod Jun 3 11:30:46.484: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6086 exec execpod-affinitydc5pb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-timeout 80' Jun 3 11:30:46.735: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-timeout 80\nConnection to affinity-clusterip-timeout 80 port [tcp/http] succeeded!\n" Jun 3 11:30:46.735: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:30:46.735: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6086 exec execpod-affinitydc5pb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.128.208 80' Jun 3 11:30:46.982: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.128.208 80\nConnection to 10.96.128.208 80 port [tcp/http] succeeded!\n" Jun 3 11:30:46.982: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:30:46.982: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6086 exec execpod-affinitydc5pb -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.96.128.208:80/ ; done' Jun 3 11:30:47.351: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n" Jun 3 11:30:47.351: INFO: stdout: "\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k\naffinity-clusterip-timeout-sqb6k" Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Received response from host: affinity-clusterip-timeout-sqb6k Jun 3 11:30:47.351: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6086 exec execpod-affinitydc5pb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.96.128.208:80/' Jun 3 11:30:47.608: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n" Jun 3 11:30:47.609: INFO: stdout: "affinity-clusterip-timeout-sqb6k" Jun 3 11:31:07.609: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6086 exec execpod-affinitydc5pb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.96.128.208:80/' Jun 3 11:31:07.847: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n" Jun 3 11:31:07.847: INFO: stdout: "affinity-clusterip-timeout-sqb6k" Jun 3 11:31:27.848: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6086 exec execpod-affinitydc5pb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.96.128.208:80/' Jun 3 11:31:28.068: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.96.128.208:80/\n" Jun 3 11:31:28.068: INFO: stdout: "affinity-clusterip-timeout-hm2v6" Jun 3 11:31:28.068: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-clusterip-timeout in namespace services-6086, will wait for the garbage collector to delete the pods Jun 3 11:31:28.134: INFO: Deleting ReplicationController affinity-clusterip-timeout took: 4.415695ms Jun 3 11:31:28.235: INFO: Terminating ReplicationController affinity-clusterip-timeout pods took: 100.920299ms [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:35.548: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-6086" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:58.015 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance]","total":-1,"completed":15,"skipped":301,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:28.262: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:31:28.698: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:31:30.709: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:31:32.713: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316688, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:31:35.722: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate configmap [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering the mutating configmap webhook via the AdmissionRegistration API STEP: create a configmap that should be updated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:35.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-534" for this suite. STEP: Destroying namespace "webhook-534-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:7.540 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate configmap [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":-1,"completed":34,"skipped":588,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:35.579: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:37.639: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-5059" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","total":-1,"completed":16,"skipped":311,"failed":0} [BeforeEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:37.650: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to start watching from a specific resource version [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a new configmap STEP: modifying the configmap once STEP: modifying the configmap a second time STEP: deleting the configmap STEP: creating a watch on configmaps from the resource version returned by the first update STEP: Expecting to observe notifications for all changes to the configmap after the first update Jun 3 11:31:37.703: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-2613 10ca96b4-1d2f-463f-b56c-696e4b54801c 3294114 0 2021-06-03 11:31:37 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2021-06-03 11:31:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:31:37.703: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-2613 10ca96b4-1d2f-463f-b56c-696e4b54801c 3294115 0 2021-06-03 11:31:37 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2021-06-03 11:31:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:37.704: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-2613" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","total":-1,"completed":17,"skipped":311,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:32.167: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating projection with secret that has name projected-secret-test-10fcd69f-f6ac-475f-b60e-71886f7355f2 STEP: Creating a pod to test consume secrets Jun 3 11:31:32.209: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115" in namespace "projected-3281" to be "Succeeded or Failed" Jun 3 11:31:32.212: INFO: Pod "pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115": Phase="Pending", Reason="", readiness=false. Elapsed: 3.224495ms Jun 3 11:31:34.217: INFO: Pod "pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007684501s Jun 3 11:31:36.221: INFO: Pod "pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01167461s Jun 3 11:31:38.225: INFO: Pod "pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016549856s STEP: Saw pod success Jun 3 11:31:38.226: INFO: Pod "pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115" satisfied condition "Succeeded or Failed" Jun 3 11:31:38.229: INFO: Trying to get logs from node v1.21-worker pod pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115 container projected-secret-volume-test: STEP: delete the pod Jun 3 11:31:38.245: INFO: Waiting for pod pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115 to disappear Jun 3 11:31:38.249: INFO: Pod pod-projected-secrets-19fc4c68-489f-4d50-b5cd-36ce821c4115 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:38.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3281" for this suite. • [SLOW TEST:6.092 seconds] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":16,"skipped":239,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:37.754: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0644 on node default medium Jun 3 11:31:37.791: INFO: Waiting up to 5m0s for pod "pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0" in namespace "emptydir-3420" to be "Succeeded or Failed" Jun 3 11:31:37.794: INFO: Pod "pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0": Phase="Pending", Reason="", readiness=false. Elapsed: 3.041844ms Jun 3 11:31:39.798: INFO: Pod "pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007469969s Jun 3 11:31:41.802: INFO: Pod "pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01116034s STEP: Saw pod success Jun 3 11:31:41.802: INFO: Pod "pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0" satisfied condition "Succeeded or Failed" Jun 3 11:31:41.805: INFO: Trying to get logs from node v1.21-worker pod pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0 container test-container: STEP: delete the pod Jun 3 11:31:41.821: INFO: Waiting for pod pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0 to disappear Jun 3 11:31:41.823: INFO: Pod pod-7fb33650-0095-46a4-bf5c-1c5d734b62c0 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:41.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3420" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":18,"skipped":330,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:35.821: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation and release no longer matching pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Given a Pod with a 'name' label pod-adoption-release is created Jun 3 11:31:35.862: INFO: The status of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:37.866: INFO: The status of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:39.866: INFO: The status of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:41.866: INFO: The status of Pod pod-adoption-release is Running (Ready = true) STEP: When a replicaset with a matching selector is created STEP: Then the orphan pod is adopted STEP: When the matched label of one of its pods change Jun 3 11:31:42.885: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:43.900: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-7858" for this suite. • [SLOW TEST:8.086 seconds] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation and release no longer matching pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","total":-1,"completed":35,"skipped":597,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:43.987: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should be updated [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod STEP: submitting the pod to kubernetes Jun 3 11:31:44.023: INFO: The status of Pod pod-update-8e5c0803-d549-40c9-99d8-805cbcaf2fb0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:46.027: INFO: The status of Pod pod-update-8e5c0803-d549-40c9-99d8-805cbcaf2fb0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:48.027: INFO: The status of Pod pod-update-8e5c0803-d549-40c9-99d8-805cbcaf2fb0 is Running (Ready = true) STEP: verifying the pod is in kubernetes STEP: updating the pod Jun 3 11:31:48.545: INFO: Successfully updated pod "pod-update-8e5c0803-d549-40c9-99d8-805cbcaf2fb0" STEP: verifying the updated pod is in kubernetes Jun 3 11:31:48.552: INFO: Pod update OK [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:48.552: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-7399" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Pods should be updated [NodeConformance] [Conformance]","total":-1,"completed":36,"skipped":640,"failed":0} SS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:15.651: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod with failed condition STEP: updating the pod Jun 3 11:31:16.206: INFO: Successfully updated pod "var-expansion-d2c8cd60-6036-4e80-bf9d-e8a0964ced55" STEP: waiting for pod running STEP: deleting the pod gracefully Jun 3 11:31:18.213: INFO: Deleting pod "var-expansion-d2c8cd60-6036-4e80-bf9d-e8a0964ced55" in namespace "var-expansion-6152" Jun 3 11:31:18.220: INFO: Wait up to 5m0s for pod "var-expansion-d2c8cd60-6036-4e80-bf9d-e8a0964ced55" to be fully deleted [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:50.227: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-6152" for this suite. • [SLOW TEST:154.586 seconds] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance]","total":-1,"completed":15,"skipped":220,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:50.279: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward api env vars Jun 3 11:31:50.317: INFO: Waiting up to 5m0s for pod "downward-api-be193615-2a59-4262-98a5-3fa78c834b8d" in namespace "downward-api-3960" to be "Succeeded or Failed" Jun 3 11:31:50.320: INFO: Pod "downward-api-be193615-2a59-4262-98a5-3fa78c834b8d": Phase="Pending", Reason="", readiness=false. Elapsed: 3.095869ms Jun 3 11:31:52.324: INFO: Pod "downward-api-be193615-2a59-4262-98a5-3fa78c834b8d": Phase="Running", Reason="", readiness=true. Elapsed: 2.007422002s Jun 3 11:31:54.329: INFO: Pod "downward-api-be193615-2a59-4262-98a5-3fa78c834b8d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012470494s STEP: Saw pod success Jun 3 11:31:54.329: INFO: Pod "downward-api-be193615-2a59-4262-98a5-3fa78c834b8d" satisfied condition "Succeeded or Failed" Jun 3 11:31:54.333: INFO: Trying to get logs from node v1.21-worker pod downward-api-be193615-2a59-4262-98a5-3fa78c834b8d container dapi-container: STEP: delete the pod Jun 3 11:31:54.348: INFO: Waiting for pod downward-api-be193615-2a59-4262-98a5-3fa78c834b8d to disappear Jun 3 11:31:54.354: INFO: Pod downward-api-be193615-2a59-4262-98a5-3fa78c834b8d no longer exists [AfterEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:54.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-3960" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":-1,"completed":16,"skipped":242,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:54.377: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to update and delete ResourceQuota. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a ResourceQuota STEP: Getting a ResourceQuota STEP: Updating a ResourceQuota STEP: Verifying a ResourceQuota was modified STEP: Deleting a ResourceQuota STEP: Verifying the deleted ResourceQuota [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:54.431: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-9953" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","total":-1,"completed":17,"skipped":248,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:38.293: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:52 STEP: create the container to handle the HTTPGet hook request. Jun 3 11:31:38.334: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:40.338: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:42.339: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:44.338: INFO: The status of Pod pod-handle-http-request is Running (Ready = true) [It] should execute poststart http hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the pod with lifecycle hook Jun 3 11:31:44.349: INFO: The status of Pod pod-with-poststart-http-hook is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:46.353: INFO: The status of Pod pod-with-poststart-http-hook is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:48.355: INFO: The status of Pod pod-with-poststart-http-hook is Running (Ready = true) STEP: check poststart hook STEP: delete the pod with lifecycle hook Jun 3 11:31:48.370: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jun 3 11:31:48.374: INFO: Pod pod-with-poststart-http-hook still exists Jun 3 11:31:50.374: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jun 3 11:31:50.378: INFO: Pod pod-with-poststart-http-hook still exists Jun 3 11:31:52.375: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jun 3 11:31:52.379: INFO: Pod pod-with-poststart-http-hook still exists Jun 3 11:31:54.375: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jun 3 11:31:54.379: INFO: Pod pod-with-poststart-http-hook still exists Jun 3 11:31:56.375: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jun 3 11:31:56.379: INFO: Pod pod-with-poststart-http-hook no longer exists [AfterEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:56.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-6955" for this suite. • [SLOW TEST:18.095 seconds] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:43 should execute poststart http hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","total":-1,"completed":17,"skipped":256,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:56.476: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create ConfigMap with empty key [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap that has name configmap-test-emptyKey-2373f211-75a0-4dbc-bf81-875f1b4e6534 [AfterEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:56.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-9424" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","total":-1,"completed":18,"skipped":299,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:34.570: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:31:34.612: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:36.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:38.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:40.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:42.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:44.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:46.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:48.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:50.615: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:52.615: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:54.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = false) Jun 3 11:31:56.616: INFO: The status of Pod test-webserver-eb442d29-4124-4a3d-a5d4-e1f9b2892645 is Running (Ready = true) Jun 3 11:31:56.620: INFO: Container started at 2021-06-03 11:31:36 +0000 UTC, pod became ready at 2021-06-03 11:31:54 +0000 UTC [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:56.620: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-446" for this suite. • [SLOW TEST:22.059 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","total":-1,"completed":20,"skipped":272,"failed":0} SS ------------------------------ [BeforeEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:54.524: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test override all Jun 3 11:31:54.616: INFO: Waiting up to 5m0s for pod "client-containers-2eecd690-5ab6-4ff3-9d32-7e8d3eed398d" in namespace "containers-6233" to be "Succeeded or Failed" Jun 3 11:31:54.619: INFO: Pod "client-containers-2eecd690-5ab6-4ff3-9d32-7e8d3eed398d": Phase="Pending", Reason="", readiness=false. Elapsed: 3.07232ms Jun 3 11:31:56.623: INFO: Pod "client-containers-2eecd690-5ab6-4ff3-9d32-7e8d3eed398d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007274304s STEP: Saw pod success Jun 3 11:31:56.623: INFO: Pod "client-containers-2eecd690-5ab6-4ff3-9d32-7e8d3eed398d" satisfied condition "Succeeded or Failed" Jun 3 11:31:56.626: INFO: Trying to get logs from node v1.21-worker pod client-containers-2eecd690-5ab6-4ff3-9d32-7e8d3eed398d container agnhost-container: STEP: delete the pod Jun 3 11:31:56.640: INFO: Waiting for pod client-containers-2eecd690-5ab6-4ff3-9d32-7e8d3eed398d to disappear Jun 3 11:31:56.643: INFO: Pod client-containers-2eecd690-5ab6-4ff3-9d32-7e8d3eed398d no longer exists [AfterEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:56.643: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-6233" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","total":-1,"completed":18,"skipped":295,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:56.634: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should support proxy with --port 0 [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: starting the proxy server Jun 3 11:31:56.664: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7564 proxy -p 0 --disable-filter' STEP: curling proxy /api/ output [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:56.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7564" for this suite. • ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]","total":-1,"completed":21,"skipped":274,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:56.729: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should complete a service status lifecycle [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a Service STEP: watching for the Service to be added Jun 3 11:31:56.775: INFO: Found Service test-service-44qb5 in namespace services-8376 with labels: map[test-service-static:true] & ports [{http TCP 80 {0 80 } 0}] Jun 3 11:31:56.776: INFO: Service test-service-44qb5 created STEP: Getting /status Jun 3 11:31:56.779: INFO: Service test-service-44qb5 has LoadBalancer: {[]} STEP: patching the ServiceStatus STEP: watching for the Service to be patched Jun 3 11:31:56.785: INFO: observed Service test-service-44qb5 in namespace services-8376 with annotations: map[] & LoadBalancer: {[]} Jun 3 11:31:56.785: INFO: Found Service test-service-44qb5 in namespace services-8376 with annotations: map[patchedstatus:true] & LoadBalancer: {[{203.0.113.1 []}]} Jun 3 11:31:56.785: INFO: Service test-service-44qb5 has service status patched STEP: updating the ServiceStatus Jun 3 11:31:56.996: INFO: updatedStatus.Conditions: []v1.Condition{v1.Condition{Type:"StatusUpdate", Status:"True", ObservedGeneration:0, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, Reason:"E2E", Message:"Set from e2e test"}} STEP: watching for the Service to be updated Jun 3 11:31:56.998: INFO: Observed Service test-service-44qb5 in namespace services-8376 with annotations: map[] & Conditions: {[]} Jun 3 11:31:56.998: INFO: Observed event: &Service{ObjectMeta:{test-service-44qb5 services-8376 6f034fd7-149c-4856-b21e-24ef7fd36636 3294548 0 2021-06-03 11:31:56 +0000 UTC map[test-service-static:true] map[patchedstatus:true] [] [] [{e2e.test Update v1 2021-06-03 11:31:56 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:patchedstatus":{}},"f:labels":{".":{},"f:test-service-static":{}}},"f:spec":{"f:ports":{".":{},"k:{\"port\":80,\"protocol\":\"TCP\"}":{".":{},"f:name":{},"f:port":{},"f:protocol":{},"f:targetPort":{}}},"f:sessionAffinity":{},"f:type":{}},"f:status":{"f:loadBalancer":{"f:ingress":{}}}}}]},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 80 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{},ClusterIP:10.96.32.81,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,TopologyKeys:[],IPFamilyPolicy:*SingleStack,ClusterIPs:[10.96.32.81],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{LoadBalancerIngress{IP:203.0.113.1,Hostname:,Ports:[]PortStatus{},},},},Conditions:[]Condition{},},} Jun 3 11:31:56.998: INFO: Observed event: &Service{ObjectMeta:{test-service-44qb5 services-8376 6f034fd7-149c-4856-b21e-24ef7fd36636 3294549 0 2021-06-03 11:31:56 +0000 UTC map[test-service-static:true] map[patchedstatus:true] [] [] [{e2e.test Update v1 2021-06-03 11:31:56 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:patchedstatus":{}},"f:labels":{".":{},"f:test-service-static":{}}},"f:spec":{"f:ports":{".":{},"k:{\"port\":80,\"protocol\":\"TCP\"}":{".":{},"f:name":{},"f:port":{},"f:protocol":{},"f:targetPort":{}}},"f:sessionAffinity":{},"f:type":{}}}}]},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 80 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{},ClusterIP:10.96.32.81,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,TopologyKeys:[],IPFamilyPolicy:*SingleStack,ClusterIPs:[10.96.32.81],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},} Jun 3 11:31:56.999: INFO: Found Service test-service-44qb5 in namespace services-8376 with annotations: map[patchedstatus:true] & Conditions: [{StatusUpdate True 0 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] Jun 3 11:31:56.999: INFO: Service test-service-44qb5 has service status updated STEP: patching the service STEP: watching for the Service to be patched Jun 3 11:31:57.007: INFO: observed Service test-service-44qb5 in namespace services-8376 with labels: map[test-service-static:true] Jun 3 11:31:57.008: INFO: observed Service test-service-44qb5 in namespace services-8376 with labels: map[test-service-static:true] Jun 3 11:31:57.008: INFO: observed Service test-service-44qb5 in namespace services-8376 with labels: map[test-service-static:true] Jun 3 11:31:57.008: INFO: observed Service test-service-44qb5 in namespace services-8376 with labels: map[test-service-static:true] Jun 3 11:31:57.008: INFO: Found Service test-service-44qb5 in namespace services-8376 with labels: map[test-service:patched test-service-static:true] Jun 3 11:31:57.008: INFO: Service test-service-44qb5 patched STEP: deleting the service STEP: watching for the Service to be deleted Jun 3 11:31:57.018: INFO: Observed event: ADDED Jun 3 11:31:57.018: INFO: Observed event: MODIFIED Jun 3 11:31:57.018: INFO: Observed event: MODIFIED Jun 3 11:31:57.018: INFO: Observed event: MODIFIED Jun 3 11:31:57.018: INFO: Observed event: MODIFIED Jun 3 11:31:57.018: INFO: Found Service test-service-44qb5 in namespace services-8376 with labels: map[test-service:patched test-service-static:true] & annotations: map[patchedstatus:true] Jun 3 11:31:57.018: INFO: Service test-service-44qb5 deleted [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:31:57.018: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8376" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • ------------------------------ {"msg":"PASSED [sig-network] Services should complete a service status lifecycle [Conformance]","total":-1,"completed":19,"skipped":337,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:04.936: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:63 W0603 11:30:04.967217 24 warnings.go:70] batch/v1beta1 CronJob is deprecated in v1.21+, unavailable in v1.25+; use batch/v1 CronJob [It] should schedule multiple jobs concurrently [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a cronjob STEP: Ensuring more than one job is running at a time STEP: Ensuring at least two running jobs exists by listing jobs explicitly STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:02.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-7485" for this suite. • [SLOW TEST:118.063 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should schedule multiple jobs concurrently [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should schedule multiple jobs concurrently [Conformance]","total":-1,"completed":34,"skipped":448,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:56.828: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name projected-configmap-test-volume-5a328e0f-7d68-4721-a236-389986bec5de STEP: Creating a pod to test consume configMaps Jun 3 11:31:57.009: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223" in namespace "projected-7238" to be "Succeeded or Failed" Jun 3 11:31:57.013: INFO: Pod "pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223": Phase="Pending", Reason="", readiness=false. Elapsed: 3.27411ms Jun 3 11:31:59.081: INFO: Pod "pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223": Phase="Pending", Reason="", readiness=false. Elapsed: 2.072167931s Jun 3 11:32:01.086: INFO: Pod "pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223": Phase="Pending", Reason="", readiness=false. Elapsed: 4.077089482s Jun 3 11:32:03.089: INFO: Pod "pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.080038744s STEP: Saw pod success Jun 3 11:32:03.089: INFO: Pod "pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223" satisfied condition "Succeeded or Failed" Jun 3 11:32:03.092: INFO: Trying to get logs from node v1.21-worker pod pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223 container agnhost-container: STEP: delete the pod Jun 3 11:32:03.107: INFO: Waiting for pod pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223 to disappear Jun 3 11:32:03.110: INFO: Pod pod-projected-configmaps-5e47a754-679d-4075-af42-6b103dbdb223 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:03.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7238" for this suite. • [SLOW TEST:6.291 seconds] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":-1,"completed":22,"skipped":301,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:56.545: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should update annotations on modification [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating the pod Jun 3 11:31:56.588: INFO: The status of Pod annotationupdateec887349-2c38-4752-91ef-006d3f174c7b is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:31:58.592: INFO: The status of Pod annotationupdateec887349-2c38-4752-91ef-006d3f174c7b is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:00.593: INFO: The status of Pod annotationupdateec887349-2c38-4752-91ef-006d3f174c7b is Running (Ready = true) Jun 3 11:32:01.114: INFO: Successfully updated pod "annotationupdateec887349-2c38-4752-91ef-006d3f174c7b" [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:03.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-8506" for this suite. • [SLOW TEST:6.594 seconds] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should update annotations on modification [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":-1,"completed":19,"skipped":311,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-instrumentation] Events API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:03.139: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-instrumentation] Events API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/events.go:81 [It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a test event STEP: listing events in all namespaces STEP: listing events in test namespace STEP: listing events with field selection filtering on source STEP: listing events with field selection filtering on reportingController STEP: getting the test event STEP: patching the test event STEP: getting the test event STEP: updating the test event STEP: getting the test event STEP: deleting the test event STEP: listing events in all namespaces STEP: listing events in test namespace [AfterEach] [sig-instrumentation] Events API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:03.221: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-5641" for this suite. • ------------------------------ {"msg":"PASSED [sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance]","total":-1,"completed":23,"skipped":311,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/sysctl.go:35 [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:03.010: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sysctl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/sysctl.go:64 [It] should support sysctls [MinimumKubeletVersion:1.21] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod with the kernel.shm_rmid_forced sysctl STEP: Watching for error events or started pod STEP: Waiting for pod completion STEP: Checking that the pod succeeded STEP: Getting logs from the pod STEP: Checking that the sysctl is actually updated [AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:07.070: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sysctl-7302" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] should support sysctls [MinimumKubeletVersion:1.21] [Conformance]","total":-1,"completed":35,"skipped":452,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:03.299: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46 [It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:32:03.337: INFO: Waiting up to 5m0s for pod "busybox-user-65534-8efafe21-1b81-46b5-9bc6-1e22f491c672" in namespace "security-context-test-8601" to be "Succeeded or Failed" Jun 3 11:32:03.340: INFO: Pod "busybox-user-65534-8efafe21-1b81-46b5-9bc6-1e22f491c672": Phase="Pending", Reason="", readiness=false. Elapsed: 3.101749ms Jun 3 11:32:05.345: INFO: Pod "busybox-user-65534-8efafe21-1b81-46b5-9bc6-1e22f491c672": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007366853s Jun 3 11:32:07.349: INFO: Pod "busybox-user-65534-8efafe21-1b81-46b5-9bc6-1e22f491c672": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01186296s Jun 3 11:32:07.349: INFO: Pod "busybox-user-65534-8efafe21-1b81-46b5-9bc6-1e22f491c672" satisfied condition "Succeeded or Failed" [AfterEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:07.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-8601" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":24,"skipped":341,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:57.061: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:31:57.756: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created Jun 3 11:31:59.767: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:32:01.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316717, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:32:04.778: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny custom resource creation, update and deletion [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:32:04.782: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the custom resource webhook via the AdmissionRegistration API STEP: Creating a custom resource that should be denied by the webhook STEP: Creating a custom resource whose deletion would be denied by the webhook STEP: Updating the custom resource with disallowed data should be denied STEP: Deleting the custom resource should be denied STEP: Remove the offending key and value from the custom resource data STEP: Deleting the updated custom resource should be successful [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:07.888: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-3422" for this suite. STEP: Destroying namespace "webhook-3422-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:10.865 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny custom resource creation, update and deletion [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":-1,"completed":20,"skipped":352,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:07.113: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-5f503947-9235-4e70-9fad-8c9ab1efbff2 STEP: Creating a pod to test consume secrets Jun 3 11:32:07.156: INFO: Waiting up to 5m0s for pod "pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a" in namespace "secrets-9457" to be "Succeeded or Failed" Jun 3 11:32:07.159: INFO: Pod "pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.766476ms Jun 3 11:32:09.163: INFO: Pod "pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006582749s Jun 3 11:32:11.167: INFO: Pod "pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01120635s Jun 3 11:32:13.171: INFO: Pod "pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015087517s Jun 3 11:32:15.175: INFO: Pod "pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.01901286s STEP: Saw pod success Jun 3 11:32:15.175: INFO: Pod "pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a" satisfied condition "Succeeded or Failed" Jun 3 11:32:15.179: INFO: Trying to get logs from node v1.21-worker pod pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a container secret-volume-test: STEP: delete the pod Jun 3 11:32:15.194: INFO: Waiting for pod pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a to disappear Jun 3 11:32:15.200: INFO: Pod pod-secrets-40a85010-7fdc-4e6d-90e7-8397f126307a no longer exists [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:15.200: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-9457" for this suite. • [SLOW TEST:8.096 seconds] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":36,"skipped":468,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:15.232: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:38 [BeforeEach] when scheduling a busybox command that always fails in a pod /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:82 [It] should be possible to delete [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-node] Kubelet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:15.278: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-2817" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":-1,"completed":37,"skipped":478,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:48.567: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a configMap. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ConfigMap STEP: Ensuring resource quota status captures configMap creation STEP: Deleting a ConfigMap STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:16.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-7187" for this suite. • [SLOW TEST:28.084 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a configMap. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","total":-1,"completed":37,"skipped":642,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:16.673: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should run through a ConfigMap lifecycle [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a ConfigMap STEP: fetching the ConfigMap STEP: patching the ConfigMap STEP: listing all ConfigMaps in all namespaces with a label selector STEP: deleting the ConfigMap by collection with a label selector STEP: listing all ConfigMaps in test namespace [AfterEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:16.730: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7820" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance]","total":-1,"completed":38,"skipped":652,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:03.219: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication Jun 3 11:32:03.724: INFO: role binding webhook-auth-reader already exists STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:32:03.738: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:32:06.756: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should honor timeout [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Setting timeout (1s) shorter than webhook latency (5s) STEP: Registering slow webhook via the AdmissionRegistration API STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s) STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore STEP: Registering slow webhook via the AdmissionRegistration API STEP: Having no error when timeout is longer than webhook latency STEP: Registering slow webhook via the AdmissionRegistration API STEP: Having no error when timeout is empty (defaulted to 10s in v1) STEP: Registering slow webhook via the AdmissionRegistration API [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:18.950: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-5308" for this suite. STEP: Destroying namespace "webhook-5308-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:15.771 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should honor timeout [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","total":-1,"completed":20,"skipped":354,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:15.362: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/pods.go:186 [It] should get a host IP [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating pod Jun 3 11:32:15.401: INFO: The status of Pod pod-hostip-ff842f85-166e-4980-a5ed-a7a93f31aafd is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:17.405: INFO: The status of Pod pod-hostip-ff842f85-166e-4980-a5ed-a7a93f31aafd is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:19.420: INFO: The status of Pod pod-hostip-ff842f85-166e-4980-a5ed-a7a93f31aafd is Running (Ready = true) Jun 3 11:32:19.427: INFO: Pod pod-hostip-ff842f85-166e-4980-a5ed-a7a93f31aafd has hostIP: 172.18.0.4 [AfterEach] [sig-node] Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:19.427: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-4495" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Pods should get a host IP [NodeConformance] [Conformance]","total":-1,"completed":38,"skipped":516,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:19.471: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should create and stop a working application [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating all guestbook components Jun 3 11:32:19.503: INFO: apiVersion: v1 kind: Service metadata: name: agnhost-replica labels: app: agnhost role: replica tier: backend spec: ports: - port: 6379 selector: app: agnhost role: replica tier: backend Jun 3 11:32:19.503: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 create -f -' Jun 3 11:32:19.910: INFO: stderr: "" Jun 3 11:32:19.910: INFO: stdout: "service/agnhost-replica created\n" Jun 3 11:32:19.910: INFO: apiVersion: v1 kind: Service metadata: name: agnhost-primary labels: app: agnhost role: primary tier: backend spec: ports: - port: 6379 targetPort: 6379 selector: app: agnhost role: primary tier: backend Jun 3 11:32:19.910: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 create -f -' Jun 3 11:32:20.216: INFO: stderr: "" Jun 3 11:32:20.216: INFO: stdout: "service/agnhost-primary created\n" Jun 3 11:32:20.216: INFO: apiVersion: v1 kind: Service metadata: name: frontend labels: app: guestbook tier: frontend spec: # if your cluster supports it, uncomment the following to automatically create # an external load-balanced IP for the frontend service. # type: LoadBalancer ports: - port: 80 selector: app: guestbook tier: frontend Jun 3 11:32:20.216: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 create -f -' Jun 3 11:32:20.600: INFO: stderr: "" Jun 3 11:32:20.601: INFO: stdout: "service/frontend created\n" Jun 3 11:32:20.601: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: frontend spec: replicas: 3 selector: matchLabels: app: guestbook tier: frontend template: metadata: labels: app: guestbook tier: frontend spec: containers: - name: guestbook-frontend image: k8s.gcr.io/e2e-test-images/agnhost:2.32 args: [ "guestbook", "--backend-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 80 Jun 3 11:32:20.601: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 create -f -' Jun 3 11:32:20.913: INFO: stderr: "" Jun 3 11:32:20.913: INFO: stdout: "deployment.apps/frontend created\n" Jun 3 11:32:20.913: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: agnhost-primary spec: replicas: 1 selector: matchLabels: app: agnhost role: primary tier: backend template: metadata: labels: app: agnhost role: primary tier: backend spec: containers: - name: primary image: k8s.gcr.io/e2e-test-images/agnhost:2.32 args: [ "guestbook", "--http-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Jun 3 11:32:20.914: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 create -f -' Jun 3 11:32:21.293: INFO: stderr: "" Jun 3 11:32:21.293: INFO: stdout: "deployment.apps/agnhost-primary created\n" Jun 3 11:32:21.294: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: agnhost-replica spec: replicas: 2 selector: matchLabels: app: agnhost role: replica tier: backend template: metadata: labels: app: agnhost role: replica tier: backend spec: containers: - name: replica image: k8s.gcr.io/e2e-test-images/agnhost:2.32 args: [ "guestbook", "--replicaof", "agnhost-primary", "--http-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Jun 3 11:32:21.294: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 create -f -' Jun 3 11:32:21.612: INFO: stderr: "" Jun 3 11:32:21.612: INFO: stdout: "deployment.apps/agnhost-replica created\n" STEP: validating guestbook app Jun 3 11:32:21.612: INFO: Waiting for all frontend pods to be Running. Jun 3 11:32:26.664: INFO: Waiting for frontend to serve content. Jun 3 11:32:26.673: INFO: Trying to add a new entry to the guestbook. Jun 3 11:32:27.695: INFO: Verifying that added entry can be retrieved. STEP: using delete to clean up resources Jun 3 11:32:27.886: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 delete --grace-period=0 --force -f -' Jun 3 11:32:28.786: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:32:28.786: INFO: stdout: "service \"agnhost-replica\" force deleted\n" STEP: using delete to clean up resources Jun 3 11:32:28.786: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 delete --grace-period=0 --force -f -' Jun 3 11:32:29.194: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:32:29.194: INFO: stdout: "service \"agnhost-primary\" force deleted\n" STEP: using delete to clean up resources Jun 3 11:32:29.195: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 delete --grace-period=0 --force -f -' Jun 3 11:32:29.607: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:32:29.607: INFO: stdout: "service \"frontend\" force deleted\n" STEP: using delete to clean up resources Jun 3 11:32:29.607: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 delete --grace-period=0 --force -f -' Jun 3 11:32:29.726: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:32:29.726: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" STEP: using delete to clean up resources Jun 3 11:32:29.727: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 delete --grace-period=0 --force -f -' Jun 3 11:32:29.848: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:32:29.849: INFO: stdout: "deployment.apps \"agnhost-primary\" force deleted\n" STEP: using delete to clean up resources Jun 3 11:32:29.849: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-5818 delete --grace-period=0 --force -f -' Jun 3 11:32:29.972: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:32:29.972: INFO: stdout: "deployment.apps \"agnhost-replica\" force deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:29.972: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5818" for this suite. • [SLOW TEST:10.510 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Guestbook application /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:336 should create and stop a working application [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance]","total":-1,"completed":39,"skipped":534,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:29.993: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:32:30.021: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8739 create -f -' Jun 3 11:32:30.423: INFO: stderr: "" Jun 3 11:32:30.423: INFO: stdout: "replicationcontroller/agnhost-primary created\n" Jun 3 11:32:30.423: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8739 create -f -' Jun 3 11:32:30.826: INFO: stderr: "" Jun 3 11:32:30.826: INFO: stdout: "service/agnhost-primary created\n" STEP: Waiting for Agnhost primary to start. Jun 3 11:32:31.831: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:32:31.831: INFO: Found 0 / 1 Jun 3 11:32:32.830: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:32:32.830: INFO: Found 0 / 1 Jun 3 11:32:33.830: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:32:33.830: INFO: Found 0 / 1 Jun 3 11:32:34.830: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:32:34.830: INFO: Found 0 / 1 Jun 3 11:32:35.831: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:32:35.832: INFO: Found 0 / 1 Jun 3 11:32:36.829: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:32:36.830: INFO: Found 1 / 1 Jun 3 11:32:36.830: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Jun 3 11:32:36.832: INFO: Selector matched 1 pods for map[app:agnhost] Jun 3 11:32:36.832: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Jun 3 11:32:36.832: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8739 describe pod agnhost-primary-js486' Jun 3 11:32:36.966: INFO: stderr: "" Jun 3 11:32:36.966: INFO: stdout: "Name: agnhost-primary-js486\nNamespace: kubectl-8739\nPriority: 0\nNode: v1.21-worker/172.18.0.4\nStart Time: Thu, 03 Jun 2021 11:32:30 +0000\nLabels: app=agnhost\n role=primary\nAnnotations: k8s.v1.cni.cncf.io/network-status:\n [{\n \"name\": \"\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.244.1.91\"\n ],\n \"mac\": \"ee:6b:a8:02:f1:77\",\n \"default\": true,\n \"dns\": {}\n }]\n k8s.v1.cni.cncf.io/networks-status:\n [{\n \"name\": \"\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.244.1.91\"\n ],\n \"mac\": \"ee:6b:a8:02:f1:77\",\n \"default\": true,\n \"dns\": {}\n }]\nStatus: Running\nIP: 10.244.1.91\nIPs:\n IP: 10.244.1.91\nControlled By: ReplicationController/agnhost-primary\nContainers:\n agnhost-primary:\n Container ID: containerd://35000b84ff020cdef4e1964c29558c6a9553fedaa1487f0d6ee66d33711b75f6\n Image: k8s.gcr.io/e2e-test-images/agnhost:2.32\n Image ID: k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Thu, 03 Jun 2021 11:32:31 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-7ssfn (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n kube-api-access-7ssfn:\n Type: Projected (a volume that contains injected data from multiple sources)\n TokenExpirationSeconds: 3607\n ConfigMapName: kube-root-ca.crt\n ConfigMapOptional: \n DownwardAPI: true\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 6s default-scheduler Successfully assigned kubectl-8739/agnhost-primary-js486 to v1.21-worker\n Normal AddedInterface 5s multus Add eth0 [10.244.1.91/24]\n Normal Pulled 5s kubelet Container image \"k8s.gcr.io/e2e-test-images/agnhost:2.32\" already present on machine\n Normal Created 5s kubelet Created container agnhost-primary\n Normal Started 5s kubelet Started container agnhost-primary\n" Jun 3 11:32:36.967: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8739 describe rc agnhost-primary' Jun 3 11:32:37.117: INFO: stderr: "" Jun 3 11:32:37.117: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-8739\nSelector: app=agnhost,role=primary\nLabels: app=agnhost\n role=primary\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=primary\n Containers:\n agnhost-primary:\n Image: k8s.gcr.io/e2e-test-images/agnhost:2.32\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 7s replication-controller Created pod: agnhost-primary-js486\n" Jun 3 11:32:37.117: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8739 describe service agnhost-primary' Jun 3 11:32:37.256: INFO: stderr: "" Jun 3 11:32:37.256: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-8739\nLabels: app=agnhost\n role=primary\nAnnotations: \nSelector: app=agnhost,role=primary\nType: ClusterIP\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.96.95.10\nIPs: 10.96.95.10\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 10.244.1.91:6379\nSession Affinity: None\nEvents: \n" Jun 3 11:32:37.260: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8739 describe node v1.21-control-plane' Jun 3 11:32:37.423: INFO: stderr: "" Jun 3 11:32:37.424: INFO: stdout: "Name: v1.21-control-plane\nRoles: control-plane,master\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n ingress-ready=true\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=v1.21-control-plane\n kubernetes.io/os=linux\n node-role.kubernetes.io/control-plane=\n node-role.kubernetes.io/master=\n node.kubernetes.io/exclude-from-external-load-balancers=\nAnnotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///run/containerd/containerd.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Mon, 24 May 2021 17:23:54 +0000\nTaints: node-role.kubernetes.io/master:NoSchedule\nUnschedulable: false\nLease:\n HolderIdentity: v1.21-control-plane\n AcquireTime: \n RenewTime: Thu, 03 Jun 2021 11:32:29 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n MemoryPressure False Thu, 03 Jun 2021 11:30:28 +0000 Mon, 24 May 2021 17:23:48 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Thu, 03 Jun 2021 11:30:28 +0000 Mon, 24 May 2021 17:23:48 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Thu, 03 Jun 2021 11:30:28 +0000 Mon, 24 May 2021 17:23:48 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Thu, 03 Jun 2021 11:30:28 +0000 Mon, 24 May 2021 17:24:29 +0000 KubeletReady kubelet is posting ready status\nAddresses:\n InternalIP: 172.18.0.3\n Hostname: v1.21-control-plane\nCapacity:\n cpu: 88\n ephemeral-storage: 459602040Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 65849824Ki\n pods: 110\nAllocatable:\n cpu: 88\n ephemeral-storage: 459602040Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 65849824Ki\n pods: 110\nSystem Info:\n Machine ID: b1187601652c41a3b6c159b2e850901f\n System UUID: c02e3c8f-3b60-418f-b9cb-d607e75a042a\n Boot ID: be455131-27dd-43f1-b9be-d55ec4651321\n Kernel Version: 5.4.0-73-generic\n OS Image: Ubuntu 20.10\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: containerd://1.5.1\n Kubelet Version: v1.21.1\n Kube-Proxy Version: v1.21.1\nPodCIDR: 10.244.0.0/24\nPodCIDRs: 10.244.0.0/24\nProviderID: kind://docker/v1.21/v1.21-control-plane\nNon-terminated Pods: (14 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age\n --------- ---- ------------ ---------- --------------- ------------- ---\n kube-system coredns-558bd4d5db-6wd7s 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 41h\n kube-system coredns-558bd4d5db-nmdj4 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 6d23h\n kube-system create-loop-devs-b8n7x 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n kube-system etcd-v1.21-control-plane 100m (0%) 0 (0%) 100Mi (0%) 0 (0%) 9d\n kube-system kindnet-x82hf 100m (0%) 100m (0%) 50Mi (0%) 50Mi (0%) 9d\n kube-system kube-apiserver-v1.21-control-plane 250m (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n kube-system kube-controller-manager-v1.21-control-plane 200m (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n kube-system kube-multus-ds-w7mzq 100m (0%) 100m (0%) 50Mi (0%) 50Mi (0%) 9d\n kube-system kube-proxy-c2smh 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n kube-system kube-scheduler-v1.21-control-plane 100m (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n kube-system tune-sysctls-t9v46 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n local-path-storage local-path-provisioner-547f784dff-7qwzp 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n metallb-system speaker-8ck5w 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9d\n projectcontour envoy-lg6jb 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9d\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 1050m (1%) 200m (0%)\n memory 340Mi (0%) 440Mi (0%)\n ephemeral-storage 100Mi (0%) 0 (0%)\n hugepages-1Gi 0 (0%) 0 (0%)\n hugepages-2Mi 0 (0%) 0 (0%)\nEvents: \n" Jun 3 11:32:37.424: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8739 describe namespace kubectl-8739' Jun 3 11:32:37.562: INFO: stderr: "" Jun 3 11:32:37.562: INFO: stdout: "Name: kubectl-8739\nLabels: e2e-framework=kubectl\n e2e-run=9fe5b590-a557-4fed-9ccf-683f0f605299\n kubernetes.io/metadata.name=kubectl-8739\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:37.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8739" for this suite. • [SLOW TEST:7.579 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl describe /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1084 should check if kubectl describe prints relevant information for rc and pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance]","total":-1,"completed":40,"skipped":541,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:07.951: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Performing setup for networking test in namespace pod-network-test-4361 STEP: creating a selector STEP: Creating the service pods in kubernetes Jun 3 11:32:07.981: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Jun 3 11:32:08.002: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:10.007: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:12.007: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:14.007: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:16.007: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:18.008: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:20.006: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:22.006: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:24.007: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:26.008: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:28.082: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:30.006: INFO: The status of Pod netserver-0 is Running (Ready = true) Jun 3 11:32:30.013: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Jun 3 11:32:38.051: INFO: Setting MaxTries for pod polling to 34 for networking test based on endpoint count 2 Jun 3 11:32:38.051: INFO: Going to poll 10.244.1.78 on port 8080 at least 0 times, with a maximum of 34 tries before failing Jun 3 11:32:38.054: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.1.78:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-4361 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:32:38.054: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:32:38.200: INFO: Found all 1 expected endpoints: [netserver-0] Jun 3 11:32:38.200: INFO: Going to poll 10.244.2.49 on port 8080 at least 0 times, with a maximum of 34 tries before failing Jun 3 11:32:38.203: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.2.49:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-4361 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:32:38.203: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:32:38.324: INFO: Found all 1 expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:38.325: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-4361" for this suite. • [SLOW TEST:30.382 seconds] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/framework.go:23 Granular Checks: Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/networking.go:30 should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":21,"skipped":364,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:38.355: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46 [It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:32:38.394: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-7c992630-1cb5-486c-86fd-33ccc13ba181" in namespace "security-context-test-9274" to be "Succeeded or Failed" Jun 3 11:32:38.397: INFO: Pod "alpine-nnp-false-7c992630-1cb5-486c-86fd-33ccc13ba181": Phase="Pending", Reason="", readiness=false. Elapsed: 2.755975ms Jun 3 11:32:40.581: INFO: Pod "alpine-nnp-false-7c992630-1cb5-486c-86fd-33ccc13ba181": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.186478134s Jun 3 11:32:40.581: INFO: Pod "alpine-nnp-false-7c992630-1cb5-486c-86fd-33ccc13ba181" satisfied condition "Succeeded or Failed" [AfterEach] [sig-node] Security Context /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:40.594: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-9274" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":22,"skipped":372,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:41.094: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should run through the lifecycle of a ServiceAccount [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a ServiceAccount STEP: watching for the ServiceAccount to be added STEP: patching the ServiceAccount STEP: finding ServiceAccount in list of all ServiceAccounts (by LabelSelector) STEP: deleting the ServiceAccount [AfterEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:41.515: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-2373" for this suite. • ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance]","total":-1,"completed":23,"skipped":438,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:41.865: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-watch STEP: Waiting for a default service account to be provisioned in namespace [It] watch on custom resource definition objects [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:31:41.893: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating first CR Jun 3 11:31:44.467: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-06-03T11:31:44Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2021-06-03T11:31:44Z]] name:name1 resourceVersion:3294333 uid:4f44cea1-224c-4dce-8d31-c3a42b9d49d6] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Creating second CR Jun 3 11:31:54.473: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-06-03T11:31:54Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2021-06-03T11:31:54Z]] name:name2 resourceVersion:3294465 uid:8f6dcd03-c8f0-4159-9971-ab4df66c7a62] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Modifying first CR Jun 3 11:32:04.479: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-06-03T11:31:44Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2021-06-03T11:32:04Z]] name:name1 resourceVersion:3294809 uid:4f44cea1-224c-4dce-8d31-c3a42b9d49d6] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Modifying second CR Jun 3 11:32:14.486: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-06-03T11:31:54Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2021-06-03T11:32:14Z]] name:name2 resourceVersion:3295090 uid:8f6dcd03-c8f0-4159-9971-ab4df66c7a62] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Deleting first CR Jun 3 11:32:24.493: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-06-03T11:31:44Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2021-06-03T11:32:04Z]] name:name1 resourceVersion:3295392 uid:4f44cea1-224c-4dce-8d31-c3a42b9d49d6] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Deleting second CR Jun 3 11:32:34.500: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-06-03T11:31:54Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2021-06-03T11:32:14Z]] name:name2 resourceVersion:3295571 uid:8f6dcd03-c8f0-4159-9971-ab4df66c7a62] num:map[num1:9223372036854775807 num2:1000000]]} [AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:45.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-watch-4165" for this suite. • [SLOW TEST:63.155 seconds] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 CustomResourceDefinition Watch /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_watch.go:42 watch on custom resource definition objects [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","total":-1,"completed":19,"skipped":346,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:41.550: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should provide container's cpu request [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:32:41.687: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875" in namespace "downward-api-7492" to be "Succeeded or Failed" Jun 3 11:32:41.691: INFO: Pod "downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875": Phase="Pending", Reason="", readiness=false. Elapsed: 3.487016ms Jun 3 11:32:43.694: INFO: Pod "downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007225154s Jun 3 11:32:45.699: INFO: Pod "downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011788047s STEP: Saw pod success Jun 3 11:32:45.699: INFO: Pod "downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875" satisfied condition "Succeeded or Failed" Jun 3 11:32:45.702: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875 container client-container: STEP: delete the pod Jun 3 11:32:45.733: INFO: Waiting for pod downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875 to disappear Jun 3 11:32:45.743: INFO: Pod downwardapi-volume-9188ba3a-2b62-441a-b6b7-9c176516b875 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:45.743: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-7492" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","total":-1,"completed":24,"skipped":457,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:16.749: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Performing setup for networking test in namespace pod-network-test-3654 STEP: creating a selector STEP: Creating the service pods in kubernetes Jun 3 11:32:16.779: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Jun 3 11:32:16.802: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:18.806: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:20.806: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:22.806: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:24.806: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:26.806: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:28.878: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:30.807: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:32.806: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:34.805: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:36.806: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:38.806: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:32:40.979: INFO: The status of Pod netserver-0 is Running (Ready = true) Jun 3 11:32:40.985: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Jun 3 11:32:45.012: INFO: Setting MaxTries for pod polling to 34 for networking test based on endpoint count 2 Jun 3 11:32:45.012: INFO: Going to poll 10.244.1.80 on port 8081 at least 0 times, with a maximum of 34 tries before failing Jun 3 11:32:45.015: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.1.80 8081 | grep -v '^\s*$'] Namespace:pod-network-test-3654 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:32:45.016: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:32:46.139: INFO: Found all 1 expected endpoints: [netserver-0] Jun 3 11:32:46.139: INFO: Going to poll 10.244.2.50 on port 8081 at least 0 times, with a maximum of 34 tries before failing Jun 3 11:32:46.142: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.2.50 8081 | grep -v '^\s*$'] Namespace:pod-network-test-3654 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:32:46.142: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:32:47.229: INFO: Found all 1 expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:47.229: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-3654" for this suite. • [SLOW TEST:30.489 seconds] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/framework.go:23 Granular Checks: Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/networking.go:30 should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":39,"skipped":656,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:37.598: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD without validation schema [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:32:37.632: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Jun 3 11:32:42.260: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4899 --namespace=crd-publish-openapi-4899 create -f -' Jun 3 11:32:43.130: INFO: stderr: "" Jun 3 11:32:43.130: INFO: stdout: "e2e-test-crd-publish-openapi-1646-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" Jun 3 11:32:43.130: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4899 --namespace=crd-publish-openapi-4899 delete e2e-test-crd-publish-openapi-1646-crds test-cr' Jun 3 11:32:43.255: INFO: stderr: "" Jun 3 11:32:43.255: INFO: stdout: "e2e-test-crd-publish-openapi-1646-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" Jun 3 11:32:43.255: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4899 --namespace=crd-publish-openapi-4899 apply -f -' Jun 3 11:32:43.576: INFO: stderr: "" Jun 3 11:32:43.576: INFO: stdout: "e2e-test-crd-publish-openapi-1646-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" Jun 3 11:32:43.576: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4899 --namespace=crd-publish-openapi-4899 delete e2e-test-crd-publish-openapi-1646-crds test-cr' Jun 3 11:32:43.695: INFO: stderr: "" Jun 3 11:32:43.695: INFO: stdout: "e2e-test-crd-publish-openapi-1646-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR without validation schema Jun 3 11:32:43.695: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4899 explain e2e-test-crd-publish-openapi-1646-crds' Jun 3 11:32:43.992: INFO: stderr: "" Jun 3 11:32:43.992: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-1646-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:48.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-4899" for this suite. • [SLOW TEST:10.915 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD without validation schema [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":-1,"completed":41,"skipped":551,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:45.039: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should provide container's cpu limit [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:32:45.075: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca" in namespace "projected-9700" to be "Succeeded or Failed" Jun 3 11:32:45.078: INFO: Pod "downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca": Phase="Pending", Reason="", readiness=false. Elapsed: 2.891811ms Jun 3 11:32:47.083: INFO: Pod "downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00815125s Jun 3 11:32:49.088: INFO: Pod "downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012782999s STEP: Saw pod success Jun 3 11:32:49.088: INFO: Pod "downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca" satisfied condition "Succeeded or Failed" Jun 3 11:32:49.091: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca container client-container: STEP: delete the pod Jun 3 11:32:49.106: INFO: Waiting for pod downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca to disappear Jun 3 11:32:49.109: INFO: Pod downwardapi-volume-f4abbd0d-be59-4244-b80f-97429522acca no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:49.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9700" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","total":-1,"completed":20,"skipped":354,"failed":0} S ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:45.760: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should provide container's cpu request [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:32:45.791: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315" in namespace "projected-1836" to be "Succeeded or Failed" Jun 3 11:32:45.794: INFO: Pod "downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315": Phase="Pending", Reason="", readiness=false. Elapsed: 2.440383ms Jun 3 11:32:47.799: INFO: Pod "downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007034116s Jun 3 11:32:49.803: INFO: Pod "downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011725726s STEP: Saw pod success Jun 3 11:32:49.803: INFO: Pod "downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315" satisfied condition "Succeeded or Failed" Jun 3 11:32:49.807: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315 container client-container: STEP: delete the pod Jun 3 11:32:49.822: INFO: Waiting for pod downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315 to disappear Jun 3 11:32:49.826: INFO: Pod downwardapi-volume-d1413cc7-d273-43d7-b25c-e1d1b3e43315 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:49.826: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1836" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","total":-1,"completed":25,"skipped":462,"failed":0} SSS ------------------------------ [BeforeEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:49.843: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Jun 3 11:32:51.901: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [sig-node] Container Runtime /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:51.913: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-4942" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":-1,"completed":26,"skipped":465,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:47.251: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating projection with secret that has name projected-secret-test-9b2161b5-a3dd-4961-b0af-aeb4c5437947 STEP: Creating a pod to test consume secrets Jun 3 11:32:47.291: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f" in namespace "projected-1105" to be "Succeeded or Failed" Jun 3 11:32:47.294: INFO: Pod "pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.774064ms Jun 3 11:32:49.298: INFO: Pod "pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007619528s Jun 3 11:32:51.303: INFO: Pod "pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012077024s Jun 3 11:32:53.308: INFO: Pod "pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.017074202s STEP: Saw pod success Jun 3 11:32:53.308: INFO: Pod "pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f" satisfied condition "Succeeded or Failed" Jun 3 11:32:53.311: INFO: Trying to get logs from node v1.21-worker pod pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f container projected-secret-volume-test: STEP: delete the pod Jun 3 11:32:53.384: INFO: Waiting for pod pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f to disappear Jun 3 11:32:53.388: INFO: Pod pod-projected-secrets-3d56c28f-f0b8-45dd-825d-f8bcecfae82f no longer exists [AfterEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:53.388: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1105" for this suite. • [SLOW TEST:6.146 seconds] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":40,"skipped":664,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:49.121: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-map-16471172-7a8b-40e5-98c1-87a0a3684ae4 STEP: Creating a pod to test consume secrets Jun 3 11:32:49.162: INFO: Waiting up to 5m0s for pod "pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563" in namespace "secrets-33" to be "Succeeded or Failed" Jun 3 11:32:49.165: INFO: Pod "pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563": Phase="Pending", Reason="", readiness=false. Elapsed: 2.845003ms Jun 3 11:32:51.170: INFO: Pod "pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008603322s Jun 3 11:32:53.175: INFO: Pod "pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013106638s Jun 3 11:32:55.180: INFO: Pod "pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018127463s STEP: Saw pod success Jun 3 11:32:55.180: INFO: Pod "pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563" satisfied condition "Succeeded or Failed" Jun 3 11:32:55.183: INFO: Trying to get logs from node v1.21-worker pod pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563 container secret-volume-test: STEP: delete the pod Jun 3 11:32:55.198: INFO: Waiting for pod pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563 to disappear Jun 3 11:32:55.201: INFO: Pod pod-secrets-cc683897-9cd9-4bf6-8890-403755e66563 no longer exists [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:55.201: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-33" for this suite. • [SLOW TEST:6.088 seconds] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":21,"skipped":355,"failed":0} S ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:48.528: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [BeforeEach] Update Demo /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:293 [It] should create and stop a replication controller [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a replication controller Jun 3 11:32:48.565: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 create -f -' Jun 3 11:32:48.971: INFO: stderr: "" Jun 3 11:32:48.971: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Jun 3 11:32:48.971: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:32:49.092: INFO: stderr: "" Jun 3 11:32:49.093: INFO: stdout: "update-demo-nautilus-w7c5v update-demo-nautilus-x6mqc " Jun 3 11:32:49.093: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods update-demo-nautilus-w7c5v -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:32:49.212: INFO: stderr: "" Jun 3 11:32:49.212: INFO: stdout: "" Jun 3 11:32:49.212: INFO: update-demo-nautilus-w7c5v is created but not running Jun 3 11:32:54.215: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:32:54.344: INFO: stderr: "" Jun 3 11:32:54.344: INFO: stdout: "update-demo-nautilus-w7c5v update-demo-nautilus-x6mqc " Jun 3 11:32:54.344: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods update-demo-nautilus-w7c5v -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:32:54.462: INFO: stderr: "" Jun 3 11:32:54.462: INFO: stdout: "true" Jun 3 11:32:54.462: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods update-demo-nautilus-w7c5v -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' Jun 3 11:32:54.576: INFO: stderr: "" Jun 3 11:32:54.576: INFO: stdout: "k8s.gcr.io/e2e-test-images/nautilus:1.4" Jun 3 11:32:54.576: INFO: validating pod update-demo-nautilus-w7c5v Jun 3 11:32:54.581: INFO: got data: { "image": "nautilus.jpg" } Jun 3 11:32:54.581: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jun 3 11:32:54.581: INFO: update-demo-nautilus-w7c5v is verified up and running Jun 3 11:32:54.581: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods update-demo-nautilus-x6mqc -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:32:54.705: INFO: stderr: "" Jun 3 11:32:54.705: INFO: stdout: "true" Jun 3 11:32:54.705: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods update-demo-nautilus-x6mqc -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' Jun 3 11:32:54.826: INFO: stderr: "" Jun 3 11:32:54.826: INFO: stdout: "k8s.gcr.io/e2e-test-images/nautilus:1.4" Jun 3 11:32:54.826: INFO: validating pod update-demo-nautilus-x6mqc Jun 3 11:32:54.831: INFO: got data: { "image": "nautilus.jpg" } Jun 3 11:32:54.831: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jun 3 11:32:54.831: INFO: update-demo-nautilus-x6mqc is verified up and running STEP: using delete to clean up resources Jun 3 11:32:54.831: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 delete --grace-period=0 --force -f -' Jun 3 11:32:54.959: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:32:54.959: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Jun 3 11:32:54.959: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get rc,svc -l name=update-demo --no-headers' Jun 3 11:32:55.088: INFO: stderr: "No resources found in kubectl-1301 namespace.\n" Jun 3 11:32:55.088: INFO: stdout: "" Jun 3 11:32:55.088: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-1301 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Jun 3 11:32:55.226: INFO: stderr: "" Jun 3 11:32:55.226: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:55.226: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-1301" for this suite. • [SLOW TEST:6.706 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Update Demo /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:291 should create and stop a replication controller [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]","total":-1,"completed":42,"skipped":558,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:51.958: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] Replicaset should have a working scale subresource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating replica set "test-rs" that asks for more than the allowed pod quota Jun 3 11:32:52.001: INFO: Pod name sample-pod: Found 0 pods out of 1 Jun 3 11:32:57.006: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running STEP: getting scale subresource STEP: updating a scale subresource STEP: verifying the replicaset Spec.Replicas was modified STEP: Patch a scale subresource [AfterEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:57.024: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-8001" for this suite. • [SLOW TEST:5.075 seconds] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Replicaset should have a working scale subresource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet Replicaset should have a working scale subresource [Conformance]","total":-1,"completed":27,"skipped":484,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:53.430: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0644 on tmpfs Jun 3 11:32:53.469: INFO: Waiting up to 5m0s for pod "pod-9cb8827e-d031-4220-8f7b-c9607509c8bb" in namespace "emptydir-847" to be "Succeeded or Failed" Jun 3 11:32:53.472: INFO: Pod "pod-9cb8827e-d031-4220-8f7b-c9607509c8bb": Phase="Pending", Reason="", readiness=false. Elapsed: 3.167243ms Jun 3 11:32:55.476: INFO: Pod "pod-9cb8827e-d031-4220-8f7b-c9607509c8bb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007041258s Jun 3 11:32:57.481: INFO: Pod "pod-9cb8827e-d031-4220-8f7b-c9607509c8bb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011772128s STEP: Saw pod success Jun 3 11:32:57.481: INFO: Pod "pod-9cb8827e-d031-4220-8f7b-c9607509c8bb" satisfied condition "Succeeded or Failed" Jun 3 11:32:57.484: INFO: Trying to get logs from node v1.21-worker pod pod-9cb8827e-d031-4220-8f7b-c9607509c8bb container test-container: STEP: delete the pod Jun 3 11:32:57.500: INFO: Waiting for pod pod-9cb8827e-d031-4220-8f7b-c9607509c8bb to disappear Jun 3 11:32:57.504: INFO: Pod pod-9cb8827e-d031-4220-8f7b-c9607509c8bb no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:32:57.504: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-847" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":41,"skipped":681,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:55.290: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should test the lifecycle of a ReplicationController [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a ReplicationController STEP: waiting for RC to be added STEP: waiting for available Replicas STEP: patching ReplicationController STEP: waiting for RC to be modified STEP: patching ReplicationController status STEP: waiting for RC to be modified STEP: waiting for available Replicas STEP: fetching ReplicationController status STEP: patching ReplicationController scale STEP: waiting for RC to be modified STEP: waiting for ReplicationController's scale to be the max amount STEP: fetching ReplicationController; ensuring that it's patched STEP: updating ReplicationController status STEP: waiting for RC to be modified STEP: listing all ReplicationControllers STEP: checking that ReplicationController has expected values STEP: deleting ReplicationControllers by collection STEP: waiting for ReplicationController to have a DELETED watchEvent [AfterEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:05.309: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-9098" for this suite. • [SLOW TEST:10.029 seconds] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should test the lifecycle of a ReplicationController [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should test the lifecycle of a ReplicationController [Conformance]","total":-1,"completed":43,"skipped":593,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:07.399: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating projection with configMap that has name projected-configmap-test-upd-7bf452dd-315a-488a-b96c-b4aaa3a5e9e0 STEP: Creating the pod Jun 3 11:32:07.450: INFO: The status of Pod pod-projected-configmaps-d6043845-10bb-4a39-976d-de57ac0b6886 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:09.454: INFO: The status of Pod pod-projected-configmaps-d6043845-10bb-4a39-976d-de57ac0b6886 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:11.455: INFO: The status of Pod pod-projected-configmaps-d6043845-10bb-4a39-976d-de57ac0b6886 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:13.457: INFO: The status of Pod pod-projected-configmaps-d6043845-10bb-4a39-976d-de57ac0b6886 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:15.454: INFO: The status of Pod pod-projected-configmaps-d6043845-10bb-4a39-976d-de57ac0b6886 is Running (Ready = true) STEP: Updating configmap projected-configmap-test-upd-7bf452dd-315a-488a-b96c-b4aaa3a5e9e0 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:13.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2655" for this suite. • [SLOW TEST:66.436 seconds] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":25,"skipped":359,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:57.044: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:86 [It] should run the lifecycle of a Deployment [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a Deployment STEP: waiting for Deployment to be created STEP: waiting for all Replicas to be Ready Jun 3 11:32:57.080: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:32:57.080: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:32:57.085: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:32:57.085: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:32:57.095: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:32:57.095: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:32:57.109: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:32:57.109: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 and labels map[test-deployment-static:true] Jun 3 11:33:03.444: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 and labels map[test-deployment-static:true] Jun 3 11:33:03.444: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 and labels map[test-deployment-static:true] Jun 3 11:33:03.840: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 and labels map[test-deployment-static:true] STEP: patching the Deployment Jun 3 11:33:03.849: INFO: observed event type ADDED STEP: waiting for Replicas to scale Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 0 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.852: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.857: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.857: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.876: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.876: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:03.885: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:03.885: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:05.832: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:05.832: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:05.848: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 STEP: listing Deployments Jun 3 11:33:05.852: INFO: Found test-deployment with labels: map[test-deployment:patched test-deployment-static:true] STEP: updating the Deployment Jun 3 11:33:05.869: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 STEP: fetching the DeploymentStatus Jun 3 11:33:05.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:05.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:05.887: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:05.898: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:05.904: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:10.448: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:10.458: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:10.463: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:10.469: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:10.474: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] Jun 3 11:33:14.850: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] STEP: patching the DeploymentStatus STEP: fetching the DeploymentStatus Jun 3 11:33:14.877: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:14.877: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:14.877: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:14.877: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:14.877: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 1 Jun 3 11:33:14.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:14.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:14.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:14.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:14.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 2 Jun 3 11:33:14.878: INFO: observed Deployment test-deployment in namespace deployment-2262 with ReadyReplicas 3 STEP: deleting the Deployment Jun 3 11:33:14.887: INFO: observed event type MODIFIED Jun 3 11:33:14.887: INFO: observed event type MODIFIED Jun 3 11:33:14.888: INFO: observed event type MODIFIED Jun 3 11:33:14.888: INFO: observed event type MODIFIED Jun 3 11:33:14.888: INFO: observed event type MODIFIED Jun 3 11:33:14.889: INFO: observed event type MODIFIED Jun 3 11:33:14.889: INFO: observed event type MODIFIED Jun 3 11:33:14.889: INFO: observed event type MODIFIED Jun 3 11:33:14.891: INFO: observed event type MODIFIED Jun 3 11:33:14.891: INFO: observed event type MODIFIED Jun 3 11:33:14.892: INFO: observed event type MODIFIED Jun 3 11:33:14.892: INFO: observed event type MODIFIED [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:80 Jun 3 11:33:14.894: INFO: Log out all the ReplicaSets if there is no deployment created [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:14.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-2262" for this suite. • [SLOW TEST:17.864 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run the lifecycle of a Deployment [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Deployment should run the lifecycle of a Deployment [Conformance]","total":-1,"completed":28,"skipped":489,"failed":0} SSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:13.856: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should provide container's memory request [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:33:13.894: INFO: Waiting up to 5m0s for pod "downwardapi-volume-bbd058dd-fd15-4365-bec1-08d7e3026c9e" in namespace "downward-api-7808" to be "Succeeded or Failed" Jun 3 11:33:13.898: INFO: Pod "downwardapi-volume-bbd058dd-fd15-4365-bec1-08d7e3026c9e": Phase="Pending", Reason="", readiness=false. Elapsed: 3.360575ms Jun 3 11:33:15.903: INFO: Pod "downwardapi-volume-bbd058dd-fd15-4365-bec1-08d7e3026c9e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.00864232s STEP: Saw pod success Jun 3 11:33:15.903: INFO: Pod "downwardapi-volume-bbd058dd-fd15-4365-bec1-08d7e3026c9e" satisfied condition "Succeeded or Failed" Jun 3 11:33:15.907: INFO: Trying to get logs from node v1.21-worker2 pod downwardapi-volume-bbd058dd-fd15-4365-bec1-08d7e3026c9e container client-container: STEP: delete the pod Jun 3 11:33:15.920: INFO: Waiting for pod downwardapi-volume-bbd058dd-fd15-4365-bec1-08d7e3026c9e to disappear Jun 3 11:33:15.923: INFO: Pod downwardapi-volume-bbd058dd-fd15-4365-bec1-08d7e3026c9e no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:15.923: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-7808" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","total":-1,"completed":26,"skipped":368,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:29:10.097: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod test-webserver-7b2e529c-1d7e-49e4-8306-b2d1ec27f0af in namespace container-probe-3114 Jun 3 11:29:18.150: INFO: Started pod test-webserver-7b2e529c-1d7e-49e4-8306-b2d1ec27f0af in namespace container-probe-3114 STEP: checking the pod's current state and verifying that restartCount is present Jun 3 11:29:18.153: INFO: Initial restart count of pod test-webserver-7b2e529c-1d7e-49e4-8306-b2d1ec27f0af is 0 STEP: deleting the pod [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:19.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-3114" for this suite. • [SLOW TEST:249.005 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":-1,"completed":10,"skipped":186,"failed":0} SSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:28:20.306: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:63 W0603 11:28:20.509830 30 warnings.go:70] batch/v1beta1 CronJob is deprecated in v1.21+, unavailable in v1.25+; use batch/v1 CronJob [It] should not schedule jobs when suspended [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a suspended cronjob STEP: Ensuring no jobs are scheduled STEP: Ensuring no job exists by listing jobs explicitly STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:20.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-7735" for this suite. • [SLOW TEST:300.233 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should not schedule jobs when suspended [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should not schedule jobs when suspended [Slow] [Conformance]","total":-1,"completed":5,"skipped":49,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:14.925: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:86 [It] RecreateDeployment should delete old pods and create new ones [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:33:14.954: INFO: Creating deployment "test-recreate-deployment" Jun 3 11:33:14.960: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 Jun 3 11:33:14.965: INFO: deployment "test-recreate-deployment" doesn't have the required revision set Jun 3 11:33:16.974: INFO: Waiting deployment "test-recreate-deployment" to complete Jun 3 11:33:16.977: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-6cb8b65c46\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:33:18.981: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316794, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-6cb8b65c46\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:33:20.980: INFO: Triggering a new rollout for deployment "test-recreate-deployment" Jun 3 11:33:20.989: INFO: Updating deployment test-recreate-deployment Jun 3 11:33:20.989: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:80 Jun 3 11:33:21.037: INFO: Deployment "test-recreate-deployment": &Deployment{ObjectMeta:{test-recreate-deployment deployment-1463 53589f0e-97cf-463c-8484-abb07fcbdd32 3296999 2 2021-06-03 11:33:14 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2021-06-03 11:33:20 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:33:21 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0013cced8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2021-06-03 11:33:21 +0000 UTC,LastTransitionTime:2021-06-03 11:33:21 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-85d47dcb4" is progressing.,LastUpdateTime:2021-06-03 11:33:21 +0000 UTC,LastTransitionTime:2021-06-03 11:33:14 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} Jun 3 11:33:21.040: INFO: New ReplicaSet "test-recreate-deployment-85d47dcb4" of Deployment "test-recreate-deployment": &ReplicaSet{ObjectMeta:{test-recreate-deployment-85d47dcb4 deployment-1463 b1eb4b8e-5dbe-456b-b2df-73f074b94fa3 3296997 1 2021-06-03 11:33:21 +0000 UTC map[name:sample-pod-3 pod-template-hash:85d47dcb4] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 53589f0e-97cf-463c-8484-abb07fcbdd32 0xc0013cd340 0xc0013cd341}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:33:21 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"53589f0e-97cf-463c-8484-abb07fcbdd32\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 85d47dcb4,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:85d47dcb4] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0013cd3b8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:33:21.040: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": Jun 3 11:33:21.040: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-6cb8b65c46 deployment-1463 c5ce44b0-5267-4b5e-8ae4-e9674831bd05 3296988 2 2021-06-03 11:33:14 +0000 UTC map[name:sample-pod-3 pod-template-hash:6cb8b65c46] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 53589f0e-97cf-463c-8484-abb07fcbdd32 0xc0013cd247 0xc0013cd248}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:33:21 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"53589f0e-97cf-463c-8484-abb07fcbdd32\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 6cb8b65c46,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:6cb8b65c46] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.32 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0013cd2d8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:33:21.043: INFO: Pod "test-recreate-deployment-85d47dcb4-sk4dr" is not available: &Pod{ObjectMeta:{test-recreate-deployment-85d47dcb4-sk4dr test-recreate-deployment-85d47dcb4- deployment-1463 3a4e3406-2720-4fde-a5d1-abbac0ce1ddc 3296994 0 2021-06-03 11:33:21 +0000 UTC map[name:sample-pod-3 pod-template-hash:85d47dcb4] map[] [{apps/v1 ReplicaSet test-recreate-deployment-85d47dcb4 b1eb4b8e-5dbe-456b-b2df-73f074b94fa3 0xc0013cd7f0 0xc0013cd7f1}] [] [{kube-controller-manager Update v1 2021-06-03 11:33:21 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"b1eb4b8e-5dbe-456b-b2df-73f074b94fa3\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-zhn7b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhn7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:33:21 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:21.043: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-1463" for this suite. • [SLOW TEST:6.126 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RecreateDeployment should delete old pods and create new ones [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","total":-1,"completed":29,"skipped":500,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:15.953: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [BeforeEach] Kubectl replace /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1548 [It] should update a single-container pod's image [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: running the image k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 Jun 3 11:33:15.984: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-4319 run e2e-test-httpd-pod --image=k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 --labels=run=e2e-test-httpd-pod' Jun 3 11:33:16.111: INFO: stderr: "" Jun 3 11:33:16.111: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: verifying the pod e2e-test-httpd-pod is running STEP: verifying the pod e2e-test-httpd-pod was created Jun 3 11:33:21.163: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-4319 get pod e2e-test-httpd-pod -o json' Jun 3 11:33:21.279: INFO: stderr: "" Jun 3 11:33:21.279: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"annotations\": {\n \"k8s.v1.cni.cncf.io/network-status\": \"[{\\n \\\"name\\\": \\\"\\\",\\n \\\"interface\\\": \\\"eth0\\\",\\n \\\"ips\\\": [\\n \\\"10.244.2.61\\\"\\n ],\\n \\\"mac\\\": \\\"32:87:26:d6:51:3c\\\",\\n \\\"default\\\": true,\\n \\\"dns\\\": {}\\n}]\",\n \"k8s.v1.cni.cncf.io/networks-status\": \"[{\\n \\\"name\\\": \\\"\\\",\\n \\\"interface\\\": \\\"eth0\\\",\\n \\\"ips\\\": [\\n \\\"10.244.2.61\\\"\\n ],\\n \\\"mac\\\": \\\"32:87:26:d6:51:3c\\\",\\n \\\"default\\\": true,\\n \\\"dns\\\": {}\\n}]\"\n },\n \"creationTimestamp\": \"2021-06-03T11:33:16Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-4319\",\n \"resourceVersion\": \"3296850\",\n \"uid\": \"cb2d4ac3-3dac-4762-9354-6a1ee319fab1\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"k8s.gcr.io/e2e-test-images/httpd:2.4.38-1\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"kube-api-access-w6lxd\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"v1.21-worker2\",\n \"preemptionPolicy\": \"PreemptLowerPriority\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"kube-api-access-w6lxd\",\n \"projected\": {\n \"defaultMode\": 420,\n \"sources\": [\n {\n \"serviceAccountToken\": {\n \"expirationSeconds\": 3607,\n \"path\": \"token\"\n }\n },\n {\n \"configMap\": {\n \"items\": [\n {\n \"key\": \"ca.crt\",\n \"path\": \"ca.crt\"\n }\n ],\n \"name\": \"kube-root-ca.crt\"\n }\n },\n {\n \"downwardAPI\": {\n \"items\": [\n {\n \"fieldRef\": {\n \"apiVersion\": \"v1\",\n \"fieldPath\": \"metadata.namespace\"\n },\n \"path\": \"namespace\"\n }\n ]\n }\n }\n ]\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2021-06-03T11:33:16Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2021-06-03T11:33:17Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2021-06-03T11:33:17Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2021-06-03T11:33:16Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://426d5a76e4e58b7fbb7522dbaaf1b57cd99978bc6a6682780108a61ec376764d\",\n \"image\": \"k8s.gcr.io/e2e-test-images/httpd:2.4.38-1\",\n \"imageID\": \"k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"started\": true,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2021-06-03T11:33:16Z\"\n }\n }\n }\n ],\n \"hostIP\": \"172.18.0.2\",\n \"phase\": \"Running\",\n \"podIP\": \"10.244.2.61\",\n \"podIPs\": [\n {\n \"ip\": \"10.244.2.61\"\n }\n ],\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2021-06-03T11:33:16Z\"\n }\n}\n" STEP: replace the image in the pod Jun 3 11:33:21.279: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-4319 replace -f -' Jun 3 11:33:21.595: INFO: stderr: "" Jun 3 11:33:21.595: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n" STEP: verifying the pod e2e-test-httpd-pod has the right image k8s.gcr.io/e2e-test-images/busybox:1.29-1 [AfterEach] Kubectl replace /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1552 Jun 3 11:33:21.598: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-4319 delete pods e2e-test-httpd-pod' Jun 3 11:33:23.911: INFO: stderr: "" Jun 3 11:33:23.911: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:23.911: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4319" for this suite. • [SLOW TEST:7.968 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl replace /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1545 should update a single-container pod's image [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance]","total":-1,"completed":27,"skipped":377,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:19.110: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-volume-3a6e6007-98eb-4453-aee6-9f83bda5f7cd STEP: Creating a pod to test consume configMaps Jun 3 11:33:19.146: INFO: Waiting up to 5m0s for pod "pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0" in namespace "configmap-7380" to be "Succeeded or Failed" Jun 3 11:33:19.149: INFO: Pod "pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.783948ms Jun 3 11:33:21.153: INFO: Pod "pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006954139s Jun 3 11:33:23.159: INFO: Pod "pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012507924s Jun 3 11:33:25.164: INFO: Pod "pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.017019956s STEP: Saw pod success Jun 3 11:33:25.164: INFO: Pod "pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0" satisfied condition "Succeeded or Failed" Jun 3 11:33:25.167: INFO: Trying to get logs from node v1.21-worker pod pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0 container agnhost-container: STEP: delete the pod Jun 3 11:33:25.187: INFO: Waiting for pod pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0 to disappear Jun 3 11:33:25.190: INFO: Pod pod-configmaps-40bd7895-e683-4982-a5c0-9a6debe61ec0 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:25.190: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7380" for this suite. • [SLOW TEST:6.087 seconds] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":-1,"completed":11,"skipped":189,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:55.213: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: http [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Performing setup for networking test in namespace pod-network-test-56 STEP: creating a selector STEP: Creating the service pods in kubernetes Jun 3 11:32:55.242: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Jun 3 11:32:55.261: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:57.265: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:32:59.266: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:01.269: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:03.267: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:05.265: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:07.266: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:09.266: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:11.266: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:13.265: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:15.271: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:17.266: INFO: The status of Pod netserver-0 is Running (Ready = false) Jun 3 11:33:19.267: INFO: The status of Pod netserver-0 is Running (Ready = true) Jun 3 11:33:19.273: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Jun 3 11:33:25.290: INFO: Setting MaxTries for pod polling to 34 for networking test based on endpoint count 2 Jun 3 11:33:25.290: INFO: Breadth first check of 10.244.1.104 on host 172.18.0.4... Jun 3 11:33:25.293: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.117:9080/dial?request=hostname&protocol=http&host=10.244.1.104&port=8080&tries=1'] Namespace:pod-network-test-56 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:33:25.293: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:33:25.403: INFO: Waiting for responses: map[] Jun 3 11:33:25.403: INFO: reached 10.244.1.104 after 0/1 tries Jun 3 11:33:25.403: INFO: Breadth first check of 10.244.2.55 on host 172.18.0.2... Jun 3 11:33:25.406: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.117:9080/dial?request=hostname&protocol=http&host=10.244.2.55&port=8080&tries=1'] Namespace:pod-network-test-56 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:33:25.406: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:33:25.528: INFO: Waiting for responses: map[] Jun 3 11:33:25.528: INFO: reached 10.244.2.55 after 0/1 tries Jun 3 11:33:25.528: INFO: Going to retry 0 out of 2 pods.... [AfterEach] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:25.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-56" for this suite. • [SLOW TEST:30.325 seconds] [sig-network] Networking /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/framework.go:23 Granular Checks: Pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/networking.go:30 should function for intra-pod communication: http [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]","total":-1,"completed":22,"skipped":356,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:21.089: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward api env vars Jun 3 11:33:21.121: INFO: Waiting up to 5m0s for pod "downward-api-955547af-4fcc-46b0-9676-26b1b922a608" in namespace "downward-api-5216" to be "Succeeded or Failed" Jun 3 11:33:21.124: INFO: Pod "downward-api-955547af-4fcc-46b0-9676-26b1b922a608": Phase="Pending", Reason="", readiness=false. Elapsed: 2.531433ms Jun 3 11:33:23.129: INFO: Pod "downward-api-955547af-4fcc-46b0-9676-26b1b922a608": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00785286s Jun 3 11:33:25.134: INFO: Pod "downward-api-955547af-4fcc-46b0-9676-26b1b922a608": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012661273s Jun 3 11:33:27.180: INFO: Pod "downward-api-955547af-4fcc-46b0-9676-26b1b922a608": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.058997381s STEP: Saw pod success Jun 3 11:33:27.180: INFO: Pod "downward-api-955547af-4fcc-46b0-9676-26b1b922a608" satisfied condition "Succeeded or Failed" Jun 3 11:33:27.183: INFO: Trying to get logs from node v1.21-worker pod downward-api-955547af-4fcc-46b0-9676-26b1b922a608 container dapi-container: STEP: delete the pod Jun 3 11:33:27.202: INFO: Waiting for pod downward-api-955547af-4fcc-46b0-9676-26b1b922a608 to disappear Jun 3 11:33:27.204: INFO: Pod downward-api-955547af-4fcc-46b0-9676-26b1b922a608 no longer exists [AfterEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:27.205: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5216" for this suite. • [SLOW TEST:6.123 seconds] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","total":-1,"completed":30,"skipped":518,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:05.330: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod pod-subpath-test-configmap-7h6x STEP: Creating a pod to test atomic-volume-subpath Jun 3 11:33:05.378: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-7h6x" in namespace "subpath-9077" to be "Succeeded or Failed" Jun 3 11:33:05.381: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Pending", Reason="", readiness=false. Elapsed: 2.92602ms Jun 3 11:33:07.385: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 2.006959328s Jun 3 11:33:09.391: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 4.01224336s Jun 3 11:33:11.395: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 6.016150922s Jun 3 11:33:13.400: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 8.021079802s Jun 3 11:33:15.405: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 10.026076912s Jun 3 11:33:17.410: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 12.031159577s Jun 3 11:33:19.414: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 14.035882733s Jun 3 11:33:21.419: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 16.040659228s Jun 3 11:33:23.424: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 18.045362859s Jun 3 11:33:25.428: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Running", Reason="", readiness=true. Elapsed: 20.049133879s Jun 3 11:33:27.477: INFO: Pod "pod-subpath-test-configmap-7h6x": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.098941141s STEP: Saw pod success Jun 3 11:33:27.477: INFO: Pod "pod-subpath-test-configmap-7h6x" satisfied condition "Succeeded or Failed" Jun 3 11:33:27.482: INFO: Trying to get logs from node v1.21-worker2 pod pod-subpath-test-configmap-7h6x container test-container-subpath-configmap-7h6x: STEP: delete the pod Jun 3 11:33:27.581: INFO: Waiting for pod pod-subpath-test-configmap-7h6x to disappear Jun 3 11:33:27.584: INFO: Pod pod-subpath-test-configmap-7h6x no longer exists STEP: Deleting pod pod-subpath-test-configmap-7h6x Jun 3 11:33:27.584: INFO: Deleting pod "pod-subpath-test-configmap-7h6x" in namespace "subpath-9077" [AfterEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:27.587: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-9077" for this suite. • [SLOW TEST:22.265 seconds] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]","total":-1,"completed":44,"skipped":598,"failed":0} [BeforeEach] [sig-instrumentation] Events API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:27.598: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-instrumentation] Events API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/events.go:81 [It] should delete a collection of events [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Create set of events STEP: get a list of Events with a label in the current namespace STEP: delete a list of events Jun 3 11:33:27.642: INFO: requesting DeleteCollection of events STEP: check that the list of events matches the requested quantity [AfterEach] [sig-instrumentation] Events API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:27.660: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-5881" for this suite. • ------------------------------ {"msg":"PASSED [sig-instrumentation] Events API should delete a collection of events [Conformance]","total":-1,"completed":45,"skipped":598,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:23.965: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should receive events on concurrent watches in same order [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: starting a background goroutine to produce watch events STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order [AfterEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:28.971: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-7334" for this suite. • [SLOW TEST:5.108 seconds] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should receive events on concurrent watches in same order [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":-1,"completed":28,"skipped":396,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:57.534: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should be able to change the type from ClusterIP to ExternalName [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-8660 STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service STEP: creating service externalsvc in namespace services-8660 STEP: creating replication controller externalsvc in namespace services-8660 I0603 11:32:57.585291 34 runners.go:190] Created replication controller with name: externalsvc, namespace: services-8660, replica count: 2 I0603 11:33:00.636469 34 runners.go:190] externalsvc Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:33:03.637644 34 runners.go:190] externalsvc Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:33:06.638823 34 runners.go:190] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady STEP: changing the ClusterIP service to type=ExternalName Jun 3 11:33:06.654: INFO: Creating new exec pod Jun 3 11:33:10.666: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-8660 exec execpodskcg9 -- /bin/sh -x -c nslookup clusterip-service.services-8660.svc.cluster.local' Jun 3 11:33:10.948: INFO: stderr: "+ nslookup clusterip-service.services-8660.svc.cluster.local\n" Jun 3 11:33:10.949: INFO: stdout: "Server:\t\t10.96.0.10\nAddress:\t10.96.0.10#53\n\nclusterip-service.services-8660.svc.cluster.local\tcanonical name = externalsvc.services-8660.svc.cluster.local.\nName:\texternalsvc.services-8660.svc.cluster.local\nAddress: 10.96.133.99\n\n" STEP: deleting ReplicationController externalsvc in namespace services-8660, will wait for the garbage collector to delete the pods Jun 3 11:33:11.008: INFO: Deleting ReplicationController externalsvc took: 5.735121ms Jun 3 11:33:11.109: INFO: Terminating ReplicationController externalsvc pods took: 100.527712ms Jun 3 11:33:29.320: INFO: Cleaning up the ClusterIP to ExternalName test service [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:29.327: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8660" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:31.803 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should be able to change the type from ClusterIP to ExternalName [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","total":-1,"completed":42,"skipped":691,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:29.191: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:33:29.228: INFO: Waiting up to 5m0s for pod "downwardapi-volume-38e1ab68-e379-4d28-9523-7955398efd08" in namespace "downward-api-8313" to be "Succeeded or Failed" Jun 3 11:33:29.231: INFO: Pod "downwardapi-volume-38e1ab68-e379-4d28-9523-7955398efd08": Phase="Pending", Reason="", readiness=false. Elapsed: 2.876458ms Jun 3 11:33:31.236: INFO: Pod "downwardapi-volume-38e1ab68-e379-4d28-9523-7955398efd08": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.008127793s STEP: Saw pod success Jun 3 11:33:31.236: INFO: Pod "downwardapi-volume-38e1ab68-e379-4d28-9523-7955398efd08" satisfied condition "Succeeded or Failed" Jun 3 11:33:31.240: INFO: Trying to get logs from node v1.21-worker2 pod downwardapi-volume-38e1ab68-e379-4d28-9523-7955398efd08 container client-container: STEP: delete the pod Jun 3 11:33:31.257: INFO: Waiting for pod downwardapi-volume-38e1ab68-e379-4d28-9523-7955398efd08 to disappear Jun 3 11:33:31.261: INFO: Pod downwardapi-volume-38e1ab68-e379-4d28-9523-7955398efd08 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:31.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-8313" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":29,"skipped":458,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/sysctl.go:35 [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:31.336: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sysctl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/sysctl.go:64 [It] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod with one valid and two invalid sysctls [AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:31.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sysctl-7567" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance]","total":-1,"completed":30,"skipped":489,"failed":0} [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:31.388: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename endpointslice STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/endpointslice.go:49 [It] should support creating EndpointSlice API operations [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: getting /apis STEP: getting /apis/discovery.k8s.io STEP: getting /apis/discovery.k8s.iov1 STEP: creating STEP: getting STEP: listing STEP: watching Jun 3 11:33:31.441: INFO: starting watch STEP: cluster-wide listing STEP: cluster-wide watching Jun 3 11:33:31.446: INFO: starting watch STEP: patching STEP: updating Jun 3 11:33:31.459: INFO: waiting for watch events with expected annotations Jun 3 11:33:31.459: INFO: saw patched and updated annotations STEP: deleting STEP: deleting a collection [AfterEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:31.481: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "endpointslice-9796" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] EndpointSlice should support creating EndpointSlice API operations [Conformance]","total":-1,"completed":31,"skipped":489,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:30:55.933: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] should have monotonically increasing restart count [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod liveness-c8c07738-827f-409e-9170-223720399edb in namespace container-probe-8408 Jun 3 11:30:57.986: INFO: Started pod liveness-c8c07738-827f-409e-9170-223720399edb in namespace container-probe-8408 STEP: checking the pod's current state and verifying that restartCount is present Jun 3 11:30:57.989: INFO: Initial restart count of pod liveness-c8c07738-827f-409e-9170-223720399edb is 0 Jun 3 11:31:17.608: INFO: Restart count of pod container-probe-8408/liveness-c8c07738-827f-409e-9170-223720399edb is now 1 (19.618707544s elapsed) Jun 3 11:31:39.653: INFO: Restart count of pod container-probe-8408/liveness-c8c07738-827f-409e-9170-223720399edb is now 2 (41.663815318s elapsed) Jun 3 11:31:59.795: INFO: Restart count of pod container-probe-8408/liveness-c8c07738-827f-409e-9170-223720399edb is now 3 (1m1.806162056s elapsed) Jun 3 11:32:19.843: INFO: Restart count of pod container-probe-8408/liveness-c8c07738-827f-409e-9170-223720399edb is now 4 (1m21.853513312s elapsed) Jun 3 11:33:32.393: INFO: Restart count of pod container-probe-8408/liveness-c8c07738-827f-409e-9170-223720399edb is now 5 (2m34.40388134s elapsed) STEP: deleting the pod [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:32.401: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-8408" for this suite. • [SLOW TEST:156.476 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should have monotonically increasing restart count [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":-1,"completed":18,"skipped":401,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:20.583: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:52 STEP: create the container to handle the HTTPGet hook request. Jun 3 11:33:20.631: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:22.636: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:24.635: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:26.634: INFO: The status of Pod pod-handle-http-request is Running (Ready = true) [It] should execute prestop http hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the pod with lifecycle hook Jun 3 11:33:26.646: INFO: The status of Pod pod-with-prestop-http-hook is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:28.650: INFO: The status of Pod pod-with-prestop-http-hook is Running (Ready = true) STEP: delete the pod with lifecycle hook Jun 3 11:33:28.657: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jun 3 11:33:28.661: INFO: Pod pod-with-prestop-http-hook still exists Jun 3 11:33:30.662: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jun 3 11:33:30.665: INFO: Pod pod-with-prestop-http-hook still exists Jun 3 11:33:32.661: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jun 3 11:33:32.665: INFO: Pod pod-with-prestop-http-hook no longer exists STEP: check prestop hook [AfterEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:32.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-1984" for this suite. • [SLOW TEST:12.096 seconds] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:43 should execute prestop http hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":-1,"completed":6,"skipped":70,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:25.228: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0777 on node default medium Jun 3 11:33:25.265: INFO: Waiting up to 5m0s for pod "pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f" in namespace "emptydir-3804" to be "Succeeded or Failed" Jun 3 11:33:25.268: INFO: Pod "pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f": Phase="Pending", Reason="", readiness=false. Elapsed: 3.030842ms Jun 3 11:33:27.272: INFO: Pod "pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006932272s Jun 3 11:33:29.275: INFO: Pod "pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010314716s Jun 3 11:33:31.279: INFO: Pod "pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f": Phase="Pending", Reason="", readiness=false. Elapsed: 6.014202131s Jun 3 11:33:33.283: INFO: Pod "pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.018121383s STEP: Saw pod success Jun 3 11:33:33.283: INFO: Pod "pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f" satisfied condition "Succeeded or Failed" Jun 3 11:33:33.286: INFO: Trying to get logs from node v1.21-worker pod pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f container test-container: STEP: delete the pod Jun 3 11:33:33.303: INFO: Waiting for pod pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f to disappear Jun 3 11:33:33.306: INFO: Pod pod-cec3bba8-50bc-4380-88ed-01a49e3fbc6f no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:33.306: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3804" for this suite. • [SLOW TEST:8.088 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":12,"skipped":206,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:27.704: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test env composition Jun 3 11:33:27.740: INFO: Waiting up to 5m0s for pod "var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54" in namespace "var-expansion-4187" to be "Succeeded or Failed" Jun 3 11:33:27.743: INFO: Pod "var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54": Phase="Pending", Reason="", readiness=false. Elapsed: 2.969611ms Jun 3 11:33:29.766: INFO: Pod "var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025739372s Jun 3 11:33:31.769: INFO: Pod "var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54": Phase="Pending", Reason="", readiness=false. Elapsed: 4.029301467s Jun 3 11:33:33.774: INFO: Pod "var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54": Phase="Pending", Reason="", readiness=false. Elapsed: 6.033979633s Jun 3 11:33:35.778: INFO: Pod "var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.038466724s STEP: Saw pod success Jun 3 11:33:35.778: INFO: Pod "var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54" satisfied condition "Succeeded or Failed" Jun 3 11:33:35.782: INFO: Trying to get logs from node v1.21-worker pod var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54 container dapi-container: STEP: delete the pod Jun 3 11:33:35.797: INFO: Waiting for pod var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54 to disappear Jun 3 11:33:35.800: INFO: Pod var-expansion-fa1f1295-c6d1-4df2-a125-0902abde2d54 no longer exists [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:35.800: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-4187" for this suite. • [SLOW TEST:8.103 seconds] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should allow composing env vars into new env vars [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":-1,"completed":46,"skipped":613,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:31.545: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should support configurable pod DNS nameservers [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod with dnsPolicy=None and customized dnsConfig... Jun 3 11:33:31.584: INFO: Created pod &Pod{ObjectMeta:{test-dns-nameservers dns-8075 08ebd389-68c5-4b4d-949a-fc56e26e3895 3297444 0 2021-06-03 11:33:31 +0000 UTC map[] map[] [] [] [{e2e.test Update v1 2021-06-03 11:33:31 +0000 UTC FieldsV1 {"f:spec":{"f:containers":{"k:{\"name\":\"agnhost-container\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsConfig":{".":{},"f:nameservers":{},"f:searches":{}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-t7744,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost-container,Image:k8s.gcr.io/e2e-test-images/agnhost:2.32,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t7744,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 3 11:33:31.588: INFO: The status of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:33.592: INFO: The status of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:35.593: INFO: The status of Pod test-dns-nameservers is Running (Ready = true) STEP: Verifying customized DNS suffix list is configured on pod... Jun 3 11:33:35.593: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-8075 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:33:35.593: INFO: >>> kubeConfig: /root/.kube/config STEP: Verifying customized DNS server is configured on pod... Jun 3 11:33:35.725: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-8075 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} Jun 3 11:33:35.725: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:33:35.864: INFO: Deleting pod test-dns-nameservers... [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:35.872: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-8075" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":-1,"completed":32,"skipped":517,"failed":0} SSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected combined /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:32.808: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-projected-all-test-volume-e0c137b1-8724-496b-8a61-2447d61d51c7 STEP: Creating secret with name secret-projected-all-test-volume-48fdcffa-0191-4d03-9e83-e068f9b27c47 STEP: Creating a pod to test Check all projections for projected volume plugin Jun 3 11:33:32.849: INFO: Waiting up to 5m0s for pod "projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3" in namespace "projected-27" to be "Succeeded or Failed" Jun 3 11:33:32.852: INFO: Pod "projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.327744ms Jun 3 11:33:34.856: INFO: Pod "projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006165572s Jun 3 11:33:36.860: INFO: Pod "projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010418851s STEP: Saw pod success Jun 3 11:33:36.860: INFO: Pod "projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3" satisfied condition "Succeeded or Failed" Jun 3 11:33:36.863: INFO: Trying to get logs from node v1.21-worker2 pod projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3 container projected-all-volume-test: STEP: delete the pod Jun 3 11:33:36.878: INFO: Waiting for pod projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3 to disappear Jun 3 11:33:36.882: INFO: Pod projected-volume-a7e498dc-0e5d-4d6e-adf0-1da283b38fd3 no longer exists [AfterEach] [sig-storage] Projected combined /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:36.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-27" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":141,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:29.374: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test emptydir 0644 on node default medium Jun 3 11:33:29.414: INFO: Waiting up to 5m0s for pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea" in namespace "emptydir-4031" to be "Succeeded or Failed" Jun 3 11:33:29.417: INFO: Pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea": Phase="Pending", Reason="", readiness=false. Elapsed: 3.25975ms Jun 3 11:33:31.422: INFO: Pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007582563s Jun 3 11:33:33.426: INFO: Pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011936276s Jun 3 11:33:35.430: INFO: Pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015572793s Jun 3 11:33:37.433: INFO: Pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea": Phase="Pending", Reason="", readiness=false. Elapsed: 8.01907656s Jun 3 11:33:39.437: INFO: Pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.023150767s STEP: Saw pod success Jun 3 11:33:39.437: INFO: Pod "pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea" satisfied condition "Succeeded or Failed" Jun 3 11:33:39.440: INFO: Trying to get logs from node v1.21-worker pod pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea container test-container: STEP: delete the pod Jun 3 11:33:39.453: INFO: Waiting for pod pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea to disappear Jun 3 11:33:39.456: INFO: Pod pod-6a8da3b2-a381-41f8-bc7c-c57b2e8542ea no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:39.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4031" for this suite. • [SLOW TEST:10.090 seconds] [sig-storage] EmptyDir volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":43,"skipped":710,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:32.475: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [BeforeEach] Kubectl label /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1308 STEP: creating the pod Jun 3 11:33:32.504: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 create -f -' Jun 3 11:33:32.829: INFO: stderr: "" Jun 3 11:33:32.829: INFO: stdout: "pod/pause created\n" Jun 3 11:33:32.829: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] Jun 3 11:33:32.829: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-9601" to be "running and ready" Jun 3 11:33:32.832: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.382377ms Jun 3 11:33:34.836: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00638972s Jun 3 11:33:36.841: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011520262s Jun 3 11:33:38.846: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 6.016486404s Jun 3 11:33:40.850: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 8.020576039s Jun 3 11:33:40.850: INFO: Pod "pause" satisfied condition "running and ready" Jun 3 11:33:40.850: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] [It] should update the label on a resource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: adding the label testing-label with value testing-label-value to a pod Jun 3 11:33:40.850: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 label pods pause testing-label=testing-label-value' Jun 3 11:33:40.995: INFO: stderr: "" Jun 3 11:33:40.995: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod has the label testing-label with the value testing-label-value Jun 3 11:33:40.995: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 get pod pause -L testing-label' Jun 3 11:33:41.294: INFO: stderr: "" Jun 3 11:33:41.294: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 9s testing-label-value\n" STEP: removing the label testing-label of a pod Jun 3 11:33:41.294: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 label pods pause testing-label-' Jun 3 11:33:41.426: INFO: stderr: "" Jun 3 11:33:41.426: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod doesn't have the label testing-label Jun 3 11:33:41.426: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 get pod pause -L testing-label' Jun 3 11:33:41.590: INFO: stderr: "" Jun 3 11:33:41.590: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 9s \n" [AfterEach] Kubectl label /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1314 STEP: using delete to clean up resources Jun 3 11:33:41.590: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 delete --grace-period=0 --force -f -' Jun 3 11:33:41.714: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:33:41.714: INFO: stdout: "pod \"pause\" force deleted\n" Jun 3 11:33:41.715: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 get rc,svc -l name=pause --no-headers' Jun 3 11:33:41.851: INFO: stderr: "No resources found in kubectl-9601 namespace.\n" Jun 3 11:33:41.851: INFO: stdout: "" Jun 3 11:33:41.851: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-9601 get pods -l name=pause -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Jun 3 11:33:41.970: INFO: stderr: "" Jun 3 11:33:41.970: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:41.970: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9601" for this suite. • [SLOW TEST:9.503 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl label /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1306 should update the label on a resource [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance]","total":-1,"completed":19,"skipped":437,"failed":0} SS ------------------------------ [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:35.909: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should release no longer matching pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Given a ReplicationController is created STEP: When the matched label of one of its pods change Jun 3 11:33:35.939: INFO: Pod name pod-release: Found 0 pods out of 1 Jun 3 11:33:40.978: INFO: Pod name pod-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:41.995: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-7655" for this suite. • [SLOW TEST:6.094 seconds] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should release no longer matching pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","total":-1,"completed":33,"skipped":531,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:31:32.256: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105 STEP: Creating service test in namespace statefulset-8367 [It] should perform canary updates and phased rolling updates of template modifications [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a new StatefulSet Jun 3 11:31:32.298: INFO: Found 0 stateful pods, waiting for 3 Jun 3 11:31:42.304: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:31:42.304: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:31:42.304: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=false Jun 3 11:31:52.305: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:31:52.305: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:31:52.305: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Updating stateful set template: update image from k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 to k8s.gcr.io/e2e-test-images/httpd:2.4.39-1 Jun 3 11:31:52.337: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Not applying an update when the partition is greater than the number of replicas STEP: Performing a canary update Jun 3 11:32:02.372: INFO: Updating stateful set ss2 Jun 3 11:32:02.380: INFO: Waiting for Pod statefulset-8367/ss2-2 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 STEP: Restoring Pods to the correct revision when they are deleted Jun 3 11:32:12.406: INFO: Found 1 stateful pods, waiting for 3 Jun 3 11:32:22.478: INFO: Found 2 stateful pods, waiting for 3 Jun 3 11:32:32.412: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:32:32.412: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:32:32.412: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Jun 3 11:32:42.411: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:32:42.411: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:32:42.411: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Performing a phased rolling update Jun 3 11:32:42.437: INFO: Updating stateful set ss2 Jun 3 11:32:42.602: INFO: Waiting for Pod statefulset-8367/ss2-1 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 Jun 3 11:32:52.612: INFO: Waiting for Pod statefulset-8367/ss2-1 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 Jun 3 11:33:02.629: INFO: Updating stateful set ss2 Jun 3 11:33:02.637: INFO: Waiting for StatefulSet statefulset-8367/ss2 to complete update Jun 3 11:33:02.637: INFO: Waiting for Pod statefulset-8367/ss2-0 to have revision ss2-5bbbc9fc94 update revision ss2-677d6db895 [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116 Jun 3 11:33:12.645: INFO: Deleting all statefulset in ns statefulset-8367 Jun 3 11:33:12.648: INFO: Scaling statefulset ss2 to 0 Jun 3 11:33:42.682: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:33:42.686: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:42.698: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-8367" for this suite. • [SLOW TEST:130.451 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95 should perform canary updates and phased rolling updates of template modifications [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","total":-1,"completed":21,"skipped":523,"failed":0} SS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:27.229: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:33:35.271: INFO: Deleting pod "var-expansion-ed416339-b809-4b8d-b91b-6e23ce8e3c2a" in namespace "var-expansion-1938" Jun 3 11:33:35.277: INFO: Wait up to 5m0s for pod "var-expansion-ed416339-b809-4b8d-b91b-6e23ce8e3c2a" to be fully deleted [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:47.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-1938" for this suite. • [SLOW TEST:20.064 seconds] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should fail substituting values in a volume subpath with absolute path [Slow] [Conformance]","total":-1,"completed":31,"skipped":526,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:41.984: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication Jun 3 11:33:42.535: INFO: role binding webhook-auth-reader already exists STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:33:42.559: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:33:45.577: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource with different stored version [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:33:45.581: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-1705-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource while v1 is storage version STEP: Patching Custom Resource Definition to set v2 as storage STEP: Patching the custom resource while v2 is storage version [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:48.798: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-666" for this suite. STEP: Destroying namespace "webhook-666-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:6.852 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource with different stored version [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":-1,"completed":20,"skipped":439,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:39.478: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:33:40.063: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:33:42.071: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:33:44.074: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 3 11:33:46.075: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316820, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:33:49.082: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate pod and apply defaults after mutation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering the mutating pod webhook via the AdmissionRegistration API STEP: create a pod that should be updated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:49.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-7648" for this suite. STEP: Destroying namespace "webhook-7648-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:9.686 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate pod and apply defaults after mutation [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:42.713: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [BeforeEach] Kubectl logs /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1386 STEP: creating an pod Jun 3 11:33:42.743: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 run logs-generator --image=k8s.gcr.io/e2e-test-images/agnhost:2.32 --restart=Never -- logs-generator --log-lines-total 100 --run-duration 20s' Jun 3 11:33:42.901: INFO: stderr: "" Jun 3 11:33:42.901: INFO: stdout: "pod/logs-generator created\n" [It] should be able to retrieve and filter logs [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Waiting for log generator to start. Jun 3 11:33:42.901: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] Jun 3 11:33:42.901: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-7820" to be "running and ready, or succeeded" Jun 3 11:33:42.981: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 80.251126ms Jun 3 11:33:44.985: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 2.084264962s Jun 3 11:33:44.985: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" Jun 3 11:33:44.985: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] STEP: checking for a matching strings Jun 3 11:33:44.986: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 logs logs-generator logs-generator' Jun 3 11:33:45.146: INFO: stderr: "" Jun 3 11:33:45.146: INFO: stdout: "I0603 11:33:43.861516 1 logs_generator.go:76] 0 POST /api/v1/namespaces/default/pods/pvs 499\nI0603 11:33:44.061652 1 logs_generator.go:76] 1 PUT /api/v1/namespaces/default/pods/4zpf 417\nI0603 11:33:44.262346 1 logs_generator.go:76] 2 POST /api/v1/namespaces/ns/pods/hkqd 521\nI0603 11:33:44.461702 1 logs_generator.go:76] 3 POST /api/v1/namespaces/default/pods/gzgg 329\nI0603 11:33:44.662124 1 logs_generator.go:76] 4 POST /api/v1/namespaces/kube-system/pods/j6z6 436\nI0603 11:33:44.862595 1 logs_generator.go:76] 5 GET /api/v1/namespaces/default/pods/6d8b 548\nI0603 11:33:45.062060 1 logs_generator.go:76] 6 POST /api/v1/namespaces/default/pods/zrxw 365\n" STEP: limiting log lines Jun 3 11:33:45.146: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 logs logs-generator logs-generator --tail=1' Jun 3 11:33:45.275: INFO: stderr: "" Jun 3 11:33:45.275: INFO: stdout: "I0603 11:33:45.262482 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/gqx5 305\n" Jun 3 11:33:45.275: INFO: got output "I0603 11:33:45.262482 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/gqx5 305\n" STEP: limiting log bytes Jun 3 11:33:45.275: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 logs logs-generator logs-generator --limit-bytes=1' Jun 3 11:33:45.404: INFO: stderr: "" Jun 3 11:33:45.404: INFO: stdout: "I" Jun 3 11:33:45.404: INFO: got output "I" STEP: exposing timestamps Jun 3 11:33:45.405: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 logs logs-generator logs-generator --tail=1 --timestamps' Jun 3 11:33:45.538: INFO: stderr: "" Jun 3 11:33:45.538: INFO: stdout: "2021-06-03T11:33:45.461866261Z I0603 11:33:45.461707 1 logs_generator.go:76] 8 PUT /api/v1/namespaces/default/pods/lcv 209\n" Jun 3 11:33:45.538: INFO: got output "2021-06-03T11:33:45.461866261Z I0603 11:33:45.461707 1 logs_generator.go:76] 8 PUT /api/v1/namespaces/default/pods/lcv 209\n" STEP: restricting to a time range Jun 3 11:33:48.040: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 logs logs-generator logs-generator --since=1s' Jun 3 11:33:48.173: INFO: stderr: "" Jun 3 11:33:48.174: INFO: stdout: "I0603 11:33:47.262428 1 logs_generator.go:76] 17 PUT /api/v1/namespaces/ns/pods/xhv 277\nI0603 11:33:47.461801 1 logs_generator.go:76] 18 POST /api/v1/namespaces/default/pods/xh8 441\nI0603 11:33:47.662293 1 logs_generator.go:76] 19 GET /api/v1/namespaces/default/pods/jf5 260\nI0603 11:33:47.861655 1 logs_generator.go:76] 20 PUT /api/v1/namespaces/kube-system/pods/j75 211\nI0603 11:33:48.062120 1 logs_generator.go:76] 21 PUT /api/v1/namespaces/ns/pods/tn7v 233\n" Jun 3 11:33:48.174: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 logs logs-generator logs-generator --since=24h' Jun 3 11:33:48.341: INFO: stderr: "" Jun 3 11:33:48.341: INFO: stdout: "I0603 11:33:43.861516 1 logs_generator.go:76] 0 POST /api/v1/namespaces/default/pods/pvs 499\nI0603 11:33:44.061652 1 logs_generator.go:76] 1 PUT /api/v1/namespaces/default/pods/4zpf 417\nI0603 11:33:44.262346 1 logs_generator.go:76] 2 POST /api/v1/namespaces/ns/pods/hkqd 521\nI0603 11:33:44.461702 1 logs_generator.go:76] 3 POST /api/v1/namespaces/default/pods/gzgg 329\nI0603 11:33:44.662124 1 logs_generator.go:76] 4 POST /api/v1/namespaces/kube-system/pods/j6z6 436\nI0603 11:33:44.862595 1 logs_generator.go:76] 5 GET /api/v1/namespaces/default/pods/6d8b 548\nI0603 11:33:45.062060 1 logs_generator.go:76] 6 POST /api/v1/namespaces/default/pods/zrxw 365\nI0603 11:33:45.262482 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/gqx5 305\nI0603 11:33:45.461707 1 logs_generator.go:76] 8 PUT /api/v1/namespaces/default/pods/lcv 209\nI0603 11:33:45.662254 1 logs_generator.go:76] 9 POST /api/v1/namespaces/kube-system/pods/f4v 501\nI0603 11:33:45.861573 1 logs_generator.go:76] 10 PUT /api/v1/namespaces/kube-system/pods/4csh 203\nI0603 11:33:46.062007 1 logs_generator.go:76] 11 POST /api/v1/namespaces/kube-system/pods/mss 216\nI0603 11:33:46.262471 1 logs_generator.go:76] 12 PUT /api/v1/namespaces/ns/pods/js8k 565\nI0603 11:33:46.461787 1 logs_generator.go:76] 13 POST /api/v1/namespaces/ns/pods/tjk 285\nI0603 11:33:46.662259 1 logs_generator.go:76] 14 GET /api/v1/namespaces/ns/pods/5ghc 283\nI0603 11:33:46.861615 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/kube-system/pods/wkvc 401\nI0603 11:33:47.062069 1 logs_generator.go:76] 16 GET /api/v1/namespaces/default/pods/xws 323\nI0603 11:33:47.262428 1 logs_generator.go:76] 17 PUT /api/v1/namespaces/ns/pods/xhv 277\nI0603 11:33:47.461801 1 logs_generator.go:76] 18 POST /api/v1/namespaces/default/pods/xh8 441\nI0603 11:33:47.662293 1 logs_generator.go:76] 19 GET /api/v1/namespaces/default/pods/jf5 260\nI0603 11:33:47.861655 1 logs_generator.go:76] 20 PUT /api/v1/namespaces/kube-system/pods/j75 211\nI0603 11:33:48.062120 1 logs_generator.go:76] 21 PUT /api/v1/namespaces/ns/pods/tn7v 233\nI0603 11:33:48.262606 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/default/pods/w7bp 433\n" [AfterEach] Kubectl logs /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1391 Jun 3 11:33:48.341: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7820 delete pod logs-generator' Jun 3 11:33:50.065: INFO: stderr: "" Jun 3 11:33:50.066: INFO: stdout: "pod \"logs-generator\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:50.066: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7820" for this suite. • [SLOW TEST:7.362 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl logs /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1383 should be able to retrieve and filter logs [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance]","total":-1,"completed":22,"skipped":525,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:36.924: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching orphans and release non-matching pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: Orphaning one of the Job's Pods Jun 3 11:33:45.484: INFO: Successfully updated pod "adopt-release-ft65m" STEP: Checking that the Job readopts the Pod Jun 3 11:33:45.484: INFO: Waiting up to 15m0s for pod "adopt-release-ft65m" in namespace "job-1550" to be "adopted" Jun 3 11:33:45.487: INFO: Pod "adopt-release-ft65m": Phase="Running", Reason="", readiness=true. Elapsed: 3.108438ms Jun 3 11:33:47.582: INFO: Pod "adopt-release-ft65m": Phase="Running", Reason="", readiness=true. Elapsed: 2.098238155s Jun 3 11:33:47.582: INFO: Pod "adopt-release-ft65m" satisfied condition "adopted" STEP: Removing the labels from the Job's Pod Jun 3 11:33:48.094: INFO: Successfully updated pod "adopt-release-ft65m" STEP: Checking that the Job releases the Pod Jun 3 11:33:48.094: INFO: Waiting up to 15m0s for pod "adopt-release-ft65m" in namespace "job-1550" to be "released" Jun 3 11:33:48.097: INFO: Pod "adopt-release-ft65m": Phase="Running", Reason="", readiness=true. Elapsed: 2.659576ms Jun 3 11:33:50.105: INFO: Pod "adopt-release-ft65m": Phase="Running", Reason="", readiness=true. Elapsed: 2.010445474s Jun 3 11:33:50.105: INFO: Pod "adopt-release-ft65m" satisfied condition "released" [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:50.105: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-1550" for this suite. • [SLOW TEST:13.189 seconds] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching orphans and release non-matching pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","total":-1,"completed":8,"skipped":158,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] RuntimeClass /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:50.121: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename runtimeclass STEP: Waiting for a default service account to be provisioned in namespace [It] should support RuntimeClasses API operations [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: getting /apis STEP: getting /apis/node.k8s.io STEP: getting /apis/node.k8s.io/v1 STEP: creating STEP: watching Jun 3 11:33:50.169: INFO: starting watch STEP: getting STEP: listing STEP: patching STEP: updating Jun 3 11:33:50.185: INFO: waiting for watch events with expected annotations STEP: deleting STEP: deleting a collection [AfterEach] [sig-node] RuntimeClass /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:50.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "runtimeclass-7351" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] RuntimeClass should support RuntimeClasses API operations [Conformance]","total":-1,"completed":23,"skipped":551,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] PodTemplates /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:50.428: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename podtemplate STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a collection of pod templates [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Create set of pod templates Jun 3 11:33:50.462: INFO: created test-podtemplate-1 Jun 3 11:33:50.466: INFO: created test-podtemplate-2 Jun 3 11:33:50.470: INFO: created test-podtemplate-3 STEP: get a list of pod templates with a label in the current namespace STEP: delete collection of pod templates Jun 3 11:33:50.473: INFO: requesting DeleteCollection of pod templates STEP: check that the list of pod templates matches the requested quantity Jun 3 11:33:50.486: INFO: requesting list of pod templates to confirm quantity [AfterEach] [sig-node] PodTemplates /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:50.489: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "podtemplate-2439" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] PodTemplates should delete a collection of pod templates [Conformance]","total":-1,"completed":24,"skipped":687,"failed":0} S ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:33.339: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should be able to create a functioning NodePort service [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service nodeport-test with type=NodePort in namespace services-6716 STEP: creating replication controller nodeport-test in namespace services-6716 I0603 11:33:33.386245 26 runners.go:190] Created replication controller with name: nodeport-test, namespace: services-6716, replica count: 2 I0603 11:33:36.437761 26 runners.go:190] nodeport-test Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:33:39.438444 26 runners.go:190] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:33:39.438: INFO: Creating new exec pod Jun 3 11:33:48.682: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6716 exec execpod8wgbd -- /bin/sh -x -c echo hostName | nc -v -t -w 2 nodeport-test 80' Jun 3 11:33:48.933: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n" Jun 3 11:33:48.933: INFO: stdout: "" Jun 3 11:33:49.934: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6716 exec execpod8wgbd -- /bin/sh -x -c echo hostName | nc -v -t -w 2 nodeport-test 80' Jun 3 11:33:50.155: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n" Jun 3 11:33:50.155: INFO: stdout: "nodeport-test-d9x6k" Jun 3 11:33:50.155: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6716 exec execpod8wgbd -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.65.92 80' Jun 3 11:33:50.359: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.65.92 80\nConnection to 10.96.65.92 80 port [tcp/http] succeeded!\n" Jun 3 11:33:50.359: INFO: stdout: "nodeport-test-d9x6k" Jun 3 11:33:50.359: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6716 exec execpod8wgbd -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.4 30496' Jun 3 11:33:50.601: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.4 30496\nConnection to 172.18.0.4 30496 port [tcp/*] succeeded!\n" Jun 3 11:33:50.601: INFO: stdout: "nodeport-test-d9x6k" Jun 3 11:33:50.601: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-6716 exec execpod8wgbd -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.2 30496' Jun 3 11:33:50.870: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.2 30496\nConnection to 172.18.0.2 30496 port [tcp/*] succeeded!\n" Jun 3 11:33:50.870: INFO: stdout: "nodeport-test-fcx4g" [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:50.870: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-6716" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:17.541 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should be able to create a functioning NodePort service [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","total":-1,"completed":13,"skipped":215,"failed":0} SS ------------------------------ [BeforeEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:50.500: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap configmap-4033/configmap-test-6c019819-b39b-4a03-84a6-de8367b52941 STEP: Creating a pod to test consume configMaps Jun 3 11:33:50.538: INFO: Waiting up to 5m0s for pod "pod-configmaps-dd820707-4db4-4d9d-9800-5bbdbfbd6645" in namespace "configmap-4033" to be "Succeeded or Failed" Jun 3 11:33:50.540: INFO: Pod "pod-configmaps-dd820707-4db4-4d9d-9800-5bbdbfbd6645": Phase="Pending", Reason="", readiness=false. Elapsed: 2.638532ms Jun 3 11:33:52.544: INFO: Pod "pod-configmaps-dd820707-4db4-4d9d-9800-5bbdbfbd6645": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.006376371s STEP: Saw pod success Jun 3 11:33:52.544: INFO: Pod "pod-configmaps-dd820707-4db4-4d9d-9800-5bbdbfbd6645" satisfied condition "Succeeded or Failed" Jun 3 11:33:52.546: INFO: Trying to get logs from node v1.21-worker2 pod pod-configmaps-dd820707-4db4-4d9d-9800-5bbdbfbd6645 container env-test: STEP: delete the pod Jun 3 11:33:52.560: INFO: Waiting for pod pod-configmaps-dd820707-4db4-4d9d-9800-5bbdbfbd6645 to disappear Jun 3 11:33:52.563: INFO: Pod pod-configmaps-dd820707-4db4-4d9d-9800-5bbdbfbd6645 no longer exists [AfterEach] [sig-node] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:52.563: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4033" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","total":-1,"completed":25,"skipped":688,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:50.158: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/init_container.go:162 [It] should invoke init containers on a RestartNever pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod Jun 3 11:33:50.184: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:33:58.830: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-3796" for this suite. • [SLOW TEST:8.681 seconds] [sig-node] InitContainer [NodeConformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should invoke init containers on a RestartNever pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","total":-1,"completed":9,"skipped":182,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:50.887: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:33:50.920: INFO: Creating ReplicaSet my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85 Jun 3 11:33:50.931: INFO: Pod name my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85: Found 0 pods out of 1 Jun 3 11:33:55.935: INFO: Pod name my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85: Found 1 pods out of 1 Jun 3 11:33:55.935: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85" is running Jun 3 11:33:55.939: INFO: Pod "my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85-rnfx4" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:33:50 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:33:52 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:33:52 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-06-03 11:33:50 +0000 UTC Reason: Message:}]) Jun 3 11:33:55.940: INFO: Trying to dial the pod Jun 3 11:34:00.954: INFO: Controller my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85: Got expected result from replica 1 [my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85-rnfx4]: "my-hostname-basic-f3239ba7-0d97-4f6c-a5e9-c645c3724f85-rnfx4", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:00.954: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-9776" for this suite. • [SLOW TEST:10.077 seconds] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance]","total":-1,"completed":14,"skipped":217,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:47.313: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a secret. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Discovering how many secrets are in namespace by default STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Secret STEP: Ensuring resource quota status captures secret creation STEP: Deleting a secret STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:04.838: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-4033" for this suite. • [SLOW TEST:17.536 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a secret. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","total":-1,"completed":32,"skipped":536,"failed":0} SSS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:00.996: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's command [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test substitution in container's command Jun 3 11:34:01.039: INFO: Waiting up to 5m0s for pod "var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c" in namespace "var-expansion-3250" to be "Succeeded or Failed" Jun 3 11:34:01.042: INFO: Pod "var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.978308ms Jun 3 11:34:03.046: INFO: Pod "var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006805822s Jun 3 11:34:05.050: INFO: Pod "var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011017044s STEP: Saw pod success Jun 3 11:34:05.050: INFO: Pod "var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c" satisfied condition "Succeeded or Failed" Jun 3 11:34:05.053: INFO: Trying to get logs from node v1.21-worker2 pod var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c container dapi-container: STEP: delete the pod Jun 3 11:34:05.069: INFO: Waiting for pod var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c to disappear Jun 3 11:34:05.073: INFO: Pod var-expansion-6eeaf10e-9233-4fc3-b67c-2a64ec09d84c no longer exists [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:05.073: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-3250" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","total":-1,"completed":15,"skipped":232,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:52.589: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [BeforeEach] Kubectl run pod /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1514 [It] should create a pod from an image when restart is Never [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: running the image k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 Jun 3 11:33:52.618: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8253 run e2e-test-httpd-pod --restart=Never --image=k8s.gcr.io/e2e-test-images/httpd:2.4.38-1' Jun 3 11:33:52.752: INFO: stderr: "" Jun 3 11:33:52.752: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: verifying the pod e2e-test-httpd-pod was created [AfterEach] Kubectl run pod /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1518 Jun 3 11:33:52.756: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-8253 delete pods e2e-test-httpd-pod' Jun 3 11:34:05.497: INFO: stderr: "" Jun 3 11:34:05.497: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:05.497: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8253" for this suite. • [SLOW TEST:12.918 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl run pod /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1511 should create a pod from an image when restart is Never [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance]","total":-1,"completed":26,"skipped":697,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:05.522: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-volume-map-b71ee266-d762-47f8-833a-bda15b0a9a54 STEP: Creating a pod to test consume configMaps Jun 3 11:34:05.565: INFO: Waiting up to 5m0s for pod "pod-configmaps-120b4f1e-24d3-47e5-9508-982f6e49307a" in namespace "configmap-5115" to be "Succeeded or Failed" Jun 3 11:34:05.568: INFO: Pod "pod-configmaps-120b4f1e-24d3-47e5-9508-982f6e49307a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.893395ms Jun 3 11:34:07.573: INFO: Pod "pod-configmaps-120b4f1e-24d3-47e5-9508-982f6e49307a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007995266s STEP: Saw pod success Jun 3 11:34:07.573: INFO: Pod "pod-configmaps-120b4f1e-24d3-47e5-9508-982f6e49307a" satisfied condition "Succeeded or Failed" Jun 3 11:34:07.576: INFO: Trying to get logs from node v1.21-worker2 pod pod-configmaps-120b4f1e-24d3-47e5-9508-982f6e49307a container agnhost-container: STEP: delete the pod Jun 3 11:34:07.593: INFO: Waiting for pod pod-configmaps-120b4f1e-24d3-47e5-9508-982f6e49307a to disappear Jun 3 11:34:07.596: INFO: Pod pod-configmaps-120b4f1e-24d3-47e5-9508-982f6e49307a no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:07.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5115" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":27,"skipped":704,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:07.642: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name projected-configmap-test-volume-797d86e5-4ae3-4141-8965-307361beb98c STEP: Creating a pod to test consume configMaps Jun 3 11:34:07.682: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-23fd57e9-e592-4c37-9803-2cf0b4815d68" in namespace "projected-2992" to be "Succeeded or Failed" Jun 3 11:34:07.685: INFO: Pod "pod-projected-configmaps-23fd57e9-e592-4c37-9803-2cf0b4815d68": Phase="Pending", Reason="", readiness=false. Elapsed: 3.375226ms Jun 3 11:34:09.690: INFO: Pod "pod-projected-configmaps-23fd57e9-e592-4c37-9803-2cf0b4815d68": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007407814s STEP: Saw pod success Jun 3 11:34:09.690: INFO: Pod "pod-projected-configmaps-23fd57e9-e592-4c37-9803-2cf0b4815d68" satisfied condition "Succeeded or Failed" Jun 3 11:34:09.693: INFO: Trying to get logs from node v1.21-worker pod pod-projected-configmaps-23fd57e9-e592-4c37-9803-2cf0b4815d68 container projected-configmap-volume-test: STEP: delete the pod Jun 3 11:34:09.707: INFO: Waiting for pod pod-projected-configmaps-23fd57e9-e592-4c37-9803-2cf0b4815d68 to disappear Jun 3 11:34:09.711: INFO: Pod pod-projected-configmaps-23fd57e9-e592-4c37-9803-2cf0b4815d68 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:09.711: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2992" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":-1,"completed":28,"skipped":724,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:48.908: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with downward pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod pod-subpath-test-downwardapi-6trd STEP: Creating a pod to test atomic-volume-subpath Jun 3 11:33:48.956: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-6trd" in namespace "subpath-6353" to be "Succeeded or Failed" Jun 3 11:33:48.959: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Pending", Reason="", readiness=false. Elapsed: 3.130842ms Jun 3 11:33:50.963: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 2.007142775s Jun 3 11:33:52.968: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 4.012003207s Jun 3 11:33:54.973: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 6.016551219s Jun 3 11:33:56.978: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 8.021729043s Jun 3 11:33:58.984: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 10.027988336s Jun 3 11:34:00.989: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 12.032469885s Jun 3 11:34:02.993: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 14.037307588s Jun 3 11:34:04.999: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 16.043349373s Jun 3 11:34:07.004: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 18.048234735s Jun 3 11:34:09.009: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Running", Reason="", readiness=true. Elapsed: 20.052642433s Jun 3 11:34:11.014: INFO: Pod "pod-subpath-test-downwardapi-6trd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.057571297s STEP: Saw pod success Jun 3 11:34:11.014: INFO: Pod "pod-subpath-test-downwardapi-6trd" satisfied condition "Succeeded or Failed" Jun 3 11:34:11.017: INFO: Trying to get logs from node v1.21-worker2 pod pod-subpath-test-downwardapi-6trd container test-container-subpath-downwardapi-6trd: STEP: delete the pod Jun 3 11:34:11.031: INFO: Waiting for pod pod-subpath-test-downwardapi-6trd to disappear Jun 3 11:34:11.034: INFO: Pod pod-subpath-test-downwardapi-6trd no longer exists STEP: Deleting pod pod-subpath-test-downwardapi-6trd Jun 3 11:34:11.034: INFO: Deleting pod "pod-subpath-test-downwardapi-6trd" in namespace "subpath-6353" [AfterEach] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:11.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-6353" for this suite. • [SLOW TEST:22.147 seconds] [sig-storage] Subpath /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with downward pod [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]","total":-1,"completed":21,"skipped":480,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-node] Events /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:05.143: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: retrieving the pod Jun 3 11:34:09.205: INFO: &Pod{ObjectMeta:{send-events-8a18d5db-c938-4bb9-9354-89768869f9bd events-8004 65634cbc-7264-4e89-addc-ec75fd787049 3298931 0 2021-06-03 11:34:05 +0000 UTC map[name:foo time:181940721] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.149" ], "mac": "8a:0e:b9:0c:9f:80", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.149" ], "mac": "8a:0e:b9:0c:9f:80", "default": true, "dns": {} }]] [] [] [{e2e.test Update v1 2021-06-03 11:34:05 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:time":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"p\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":80,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:34:06 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:34:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.149\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vg6p5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:p,Image:k8s.gcr.io/e2e-test-images/agnhost:2.32,Command:[],Args:[serve-hostname],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:,HostPort:0,ContainerPort:80,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vg6p5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:05 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.4,PodIP:10.244.1.149,StartTime:2021-06-03 11:34:05 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:p,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:34:06 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/agnhost:2.32,ImageID:k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1,ContainerID:containerd://c8c5f11b93b776d02dd2c46e3e2acca4d53facd49288f51d55fc65030e2dbb90,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.149,},},EphemeralContainerStatuses:[]ContainerStatus{},},} STEP: checking for scheduler event about the pod Jun 3 11:34:11.212: INFO: Saw scheduler event for our pod. STEP: checking for kubelet event about the pod Jun 3 11:34:13.218: INFO: Saw kubelet event for our pod. STEP: deleting the pod [AfterEach] [sig-node] Events /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:13.224: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-8004" for this suite. • [SLOW TEST:8.094 seconds] [sig-node] Events /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23 should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance]","total":-1,"completed":16,"skipped":261,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-node] PreStop /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:04.858: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename prestop STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] PreStop /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:157 [It] should call prestop when killing a pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating server pod server in namespace prestop-3776 STEP: Waiting for pods to come up. STEP: Creating tester pod tester in namespace prestop-3776 STEP: Deleting pre-stop pod Jun 3 11:34:15.934: INFO: Saw: { "Hostname": "server", "Sent": null, "Received": { "prestop": 1 }, "Errors": null, "Log": [ "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." ], "StillContactingPeers": true } STEP: Deleting the server pod [AfterEach] [sig-node] PreStop /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:15.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "prestop-3776" for this suite. • [SLOW TEST:11.101 seconds] [sig-node] PreStop /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23 should call prestop when killing a pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] PreStop should call prestop when killing a pod [Conformance]","total":-1,"completed":33,"skipped":539,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Servers with support for Table transformation /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:15.986: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename tables STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Servers with support for Table transformation /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/table_conversion.go:47 [It] should return a 406 for a backend which does not implement metadata [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-api-machinery] Servers with support for Table transformation /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:16.022: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "tables-4684" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":-1,"completed":34,"skipped":551,"failed":0} SS ------------------------------ [BeforeEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:13.252: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide host IP as an env var [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward api env vars Jun 3 11:34:13.287: INFO: Waiting up to 5m0s for pod "downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1" in namespace "downward-api-2444" to be "Succeeded or Failed" Jun 3 11:34:13.290: INFO: Pod "downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.77314ms Jun 3 11:34:15.294: INFO: Pod "downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006700605s Jun 3 11:34:17.299: INFO: Pod "downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012311377s STEP: Saw pod success Jun 3 11:34:17.299: INFO: Pod "downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1" satisfied condition "Succeeded or Failed" Jun 3 11:34:17.302: INFO: Trying to get logs from node v1.21-worker2 pod downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1 container dapi-container: STEP: delete the pod Jun 3 11:34:17.316: INFO: Waiting for pod downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1 to disappear Jun 3 11:34:17.319: INFO: Pod downward-api-89248d55-2aa6-4acd-8dd9-5c25efd1b2e1 no longer exists [AfterEach] [sig-node] Downward API /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:17.319: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-2444" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","total":-1,"completed":17,"skipped":268,"failed":0} SSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:16.039: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-volume-6d8d8545-b93f-462c-bc06-dd5f0d417de0 STEP: Creating a pod to test consume configMaps Jun 3 11:34:16.079: INFO: Waiting up to 5m0s for pod "pod-configmaps-cda37f39-506a-4c0b-a311-38ae42fb7e8c" in namespace "configmap-3344" to be "Succeeded or Failed" Jun 3 11:34:16.082: INFO: Pod "pod-configmaps-cda37f39-506a-4c0b-a311-38ae42fb7e8c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.733825ms Jun 3 11:34:18.086: INFO: Pod "pod-configmaps-cda37f39-506a-4c0b-a311-38ae42fb7e8c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007028352s STEP: Saw pod success Jun 3 11:34:18.087: INFO: Pod "pod-configmaps-cda37f39-506a-4c0b-a311-38ae42fb7e8c" satisfied condition "Succeeded or Failed" Jun 3 11:34:18.092: INFO: Trying to get logs from node v1.21-worker2 pod pod-configmaps-cda37f39-506a-4c0b-a311-38ae42fb7e8c container agnhost-container: STEP: delete the pod Jun 3 11:34:18.109: INFO: Waiting for pod pod-configmaps-cda37f39-506a-4c0b-a311-38ae42fb7e8c to disappear Jun 3 11:34:18.115: INFO: Pod pod-configmaps-cda37f39-506a-4c0b-a311-38ae42fb7e8c no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:18.115: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-3344" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":-1,"completed":35,"skipped":553,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:18.169: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should provide secure master service [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:18.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-948" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • ------------------------------ {"msg":"PASSED [sig-network] Services should provide secure master service [Conformance]","total":-1,"completed":36,"skipped":583,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:09.737: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ReplicaSet STEP: Ensuring resource quota status captures replicaset creation STEP: Deleting a ReplicaSet STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:21.009: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-484" for this suite. • [SLOW TEST:11.279 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","total":-1,"completed":29,"skipped":732,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:21.039: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-16164ae5-96d4-4d7a-b26e-607b72dc6c07 STEP: Creating a pod to test consume secrets Jun 3 11:34:21.072: INFO: Waiting up to 5m0s for pod "pod-secrets-98f8d3c7-42ef-487f-8199-f91c60cafea1" in namespace "secrets-4882" to be "Succeeded or Failed" Jun 3 11:34:21.075: INFO: Pod "pod-secrets-98f8d3c7-42ef-487f-8199-f91c60cafea1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.308261ms Jun 3 11:34:23.080: INFO: Pod "pod-secrets-98f8d3c7-42ef-487f-8199-f91c60cafea1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.00720624s STEP: Saw pod success Jun 3 11:34:23.080: INFO: Pod "pod-secrets-98f8d3c7-42ef-487f-8199-f91c60cafea1" satisfied condition "Succeeded or Failed" Jun 3 11:34:23.082: INFO: Trying to get logs from node v1.21-worker2 pod pod-secrets-98f8d3c7-42ef-487f-8199-f91c60cafea1 container secret-volume-test: STEP: delete the pod Jun 3 11:34:23.095: INFO: Waiting for pod pod-secrets-98f8d3c7-42ef-487f-8199-f91c60cafea1 to disappear Jun 3 11:34:23.101: INFO: Pod pod-secrets-98f8d3c7-42ef-487f-8199-f91c60cafea1 no longer exists [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:23.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-4882" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","total":-1,"completed":30,"skipped":744,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:23.173: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:34:23.207: INFO: Waiting up to 5m0s for pod "downwardapi-volume-15636019-7851-48ee-8e9b-9a863ec26ad4" in namespace "projected-9146" to be "Succeeded or Failed" Jun 3 11:34:23.210: INFO: Pod "downwardapi-volume-15636019-7851-48ee-8e9b-9a863ec26ad4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.467087ms Jun 3 11:34:25.280: INFO: Pod "downwardapi-volume-15636019-7851-48ee-8e9b-9a863ec26ad4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.072980179s STEP: Saw pod success Jun 3 11:34:25.281: INFO: Pod "downwardapi-volume-15636019-7851-48ee-8e9b-9a863ec26ad4" satisfied condition "Succeeded or Failed" Jun 3 11:34:25.284: INFO: Trying to get logs from node v1.21-worker2 pod downwardapi-volume-15636019-7851-48ee-8e9b-9a863ec26ad4 container client-container: STEP: delete the pod Jun 3 11:34:25.297: INFO: Waiting for pod downwardapi-volume-15636019-7851-48ee-8e9b-9a863ec26ad4 to disappear Jun 3 11:34:25.300: INFO: Pod downwardapi-volume-15636019-7851-48ee-8e9b-9a863ec26ad4 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:25.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9146" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":31,"skipped":783,"failed":0} SSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:18.288: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:86 [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:34:18.318: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) Jun 3 11:34:18.325: INFO: Pod name sample-pod: Found 0 pods out of 1 Jun 3 11:34:23.330: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Jun 3 11:34:23.330: INFO: Creating deployment "test-rolling-update-deployment" Jun 3 11:34:23.334: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has Jun 3 11:34:23.338: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created Jun 3 11:34:25.345: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected Jun 3 11:34:25.348: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:80 Jun 3 11:34:25.357: INFO: Deployment "test-rolling-update-deployment": &Deployment{ObjectMeta:{test-rolling-update-deployment deployment-6611 6ae5086a-92f7-4895-8dd6-5b0a1f144e51 3299361 1 2021-06-03 11:34:23 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] [{e2e.test Update apps/v1 2021-06-03 11:34:23 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:34:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.32 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004ce2998 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-06-03 11:34:23 +0000 UTC,LastTransitionTime:2021-06-03 11:34:23 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-585b757574" has successfully progressed.,LastUpdateTime:2021-06-03 11:34:24 +0000 UTC,LastTransitionTime:2021-06-03 11:34:23 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Jun 3 11:34:25.361: INFO: New ReplicaSet "test-rolling-update-deployment-585b757574" of Deployment "test-rolling-update-deployment": &ReplicaSet{ObjectMeta:{test-rolling-update-deployment-585b757574 deployment-6611 0cf9d187-5a6e-46e6-a7b3-794feba5d050 3299350 1 2021-06-03 11:34:23 +0000 UTC map[name:sample-pod pod-template-hash:585b757574] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 6ae5086a-92f7-4895-8dd6-5b0a1f144e51 0xc004ce2e67 0xc004ce2e68}] [] [{kube-controller-manager Update apps/v1 2021-06-03 11:34:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"6ae5086a-92f7-4895-8dd6-5b0a1f144e51\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 585b757574,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:585b757574] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.32 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004ce2ef8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:34:25.361: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": Jun 3 11:34:25.361: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-6611 1d8fb5e7-20bb-41d5-822b-f63707957338 3299359 2 2021-06-03 11:34:18 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 6ae5086a-92f7-4895-8dd6-5b0a1f144e51 0xc004ce2d57 0xc004ce2d58}] [] [{e2e.test Update apps/v1 2021-06-03 11:34:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:34:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"6ae5086a-92f7-4895-8dd6-5b0a1f144e51\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc004ce2df8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:34:25.365: INFO: Pod "test-rolling-update-deployment-585b757574-qrvd7" is available: &Pod{ObjectMeta:{test-rolling-update-deployment-585b757574-qrvd7 test-rolling-update-deployment-585b757574- deployment-6611 a7a458ba-ecd2-43e4-a161-50ef12ce8f18 3299349 0 2021-06-03 11:34:23 +0000 UTC map[name:sample-pod pod-template-hash:585b757574] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.153" ], "mac": "66:e1:f1:b9:0b:55", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.153" ], "mac": "66:e1:f1:b9:0b:55", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet test-rolling-update-deployment-585b757574 0cf9d187-5a6e-46e6-a7b3-794feba5d050 0xc004ce3317 0xc004ce3318}] [] [{kube-controller-manager Update v1 2021-06-03 11:34:23 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"0cf9d187-5a6e-46e6-a7b3-794feba5d050\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:34:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:34:24 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.153\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-r5tb6,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:k8s.gcr.io/e2e-test-images/agnhost:2.32,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r5tb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:23 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:24 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:24 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:23 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.4,PodIP:10.244.1.153,StartTime:2021-06-03 11:34:23 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:34:24 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/agnhost:2.32,ImageID:k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1,ContainerID:containerd://ef8fae064870c36bf12bd907c8c46fb4ea3b28a87c6806f58402acb00159553f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.153,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:25.365: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-6611" for this suite. • [SLOW TEST:7.087 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":-1,"completed":37,"skipped":625,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:25.413: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:34:26.683: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:34:29.701: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should unconditionally reject operations on fail closed webhook [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API STEP: create a namespace for the webhook STEP: create a configmap should be unconditionally rejected by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:29.752: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-9657" for this suite. STEP: Destroying namespace "webhook-9657-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":-1,"completed":38,"skipped":646,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:29.840: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename endpointslice STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/endpointslice.go:49 [It] should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-network] EndpointSlice /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:29.917: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "endpointslice-8074" for this suite. • ------------------------------ {"msg":"PASSED [sig-network] EndpointSlice should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance]","total":-1,"completed":39,"skipped":674,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:29.933: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:86 [It] deployment should delete old replica sets [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:34:29.967: INFO: Pod name cleanup-pod: Found 0 pods out of 1 Jun 3 11:34:34.971: INFO: Pod name cleanup-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Jun 3 11:34:34.971: INFO: Creating deployment test-cleanup-deployment STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:80 Jun 3 11:34:34.984: INFO: Deployment "test-cleanup-deployment": &Deployment{ObjectMeta:{test-cleanup-deployment deployment-8684 42e96b99-8174-465a-8887-85b70639e5bf 3299600 1 2021-06-03 11:34:34 +0000 UTC map[name:cleanup-pod] map[] [] [] [{e2e.test Update apps/v1 2021-06-03 11:34:34 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.32 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0026ca038 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[]DeploymentCondition{},ReadyReplicas:0,CollisionCount:nil,},} Jun 3 11:34:34.988: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. Jun 3 11:34:34.988: INFO: All old ReplicaSets of Deployment "test-cleanup-deployment": Jun 3 11:34:34.988: INFO: &ReplicaSet{ObjectMeta:{test-cleanup-controller deployment-8684 c8bce68c-f2d5-4436-a65e-b9c963ab866f 3299601 1 2021-06-03 11:34:29 +0000 UTC map[name:cleanup-pod pod:httpd] map[] [{apps/v1 Deployment test-cleanup-deployment 42e96b99-8174-465a-8887-85b70639e5bf 0xc0026ca34f 0xc0026ca360}] [] [{e2e.test Update apps/v1 2021-06-03 11:34:29 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-03 11:34:34 +0000 UTC FieldsV1 {"f:metadata":{"f:ownerReferences":{".":{},"k:{\"uid\":\"42e96b99-8174-465a-8887-85b70639e5bf\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod pod:httpd] map[] [] [] []} {[] [] [{httpd k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0026ca3f8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Jun 3 11:34:34.991: INFO: Pod "test-cleanup-controller-cvwbh" is available: &Pod{ObjectMeta:{test-cleanup-controller-cvwbh test-cleanup-controller- deployment-8684 4dc2a684-d4fb-4a7c-95e2-05206246d09b 3299559 0 2021-06-03 11:34:29 +0000 UTC map[name:cleanup-pod pod:httpd] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.89" ], "mac": "8e:83:d7:ae:8d:88", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.89" ], "mac": "8e:83:d7:ae:8d:88", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet test-cleanup-controller c8bce68c-f2d5-4436-a65e-b9c963ab866f 0xc0026ca6b7 0xc0026ca6b8}] [] [{kube-controller-manager Update v1 2021-06-03 11:34:29 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c8bce68c-f2d5-4436-a65e-b9c963ab866f\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-03 11:34:30 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-03 11:34:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.89\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-89kkh,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-89kkh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:v1.21-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:29 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:31 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:31 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-03 11:34:29 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.2,PodIP:10.244.2.89,StartTime:2021-06-03 11:34:29 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-03 11:34:30 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/httpd:2.4.38-1,ImageID:k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50,ContainerID:containerd://419143d05446414f598e8b024ae4b92fcc906ee28f985c929ec680bf083fb8aa,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.89,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:34.991: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-8684" for this suite. • [SLOW TEST:5.066 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should delete old replica sets [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":-1,"completed":40,"skipped":678,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:35.040: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be immutable if `immutable` field is set [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:35.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8255" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be immutable if `immutable` field is set [Conformance]","total":-1,"completed":41,"skipped":699,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:35.853: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service in namespace services-3978 STEP: creating service affinity-clusterip-transition in namespace services-3978 STEP: creating replication controller affinity-clusterip-transition in namespace services-3978 I0603 11:33:35.893283 24 runners.go:190] Created replication controller with name: affinity-clusterip-transition, namespace: services-3978, replica count: 3 I0603 11:33:38.945560 24 runners.go:190] affinity-clusterip-transition Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:33:41.946696 24 runners.go:190] affinity-clusterip-transition Pods: 3 out of 3 created, 1 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:33:44.947343 24 runners.go:190] affinity-clusterip-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:33:44.958: INFO: Creating new exec pod Jun 3 11:33:49.970: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3978 exec execpod-affinityvhftz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-transition 80' Jun 3 11:33:50.182: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-transition 80\nConnection to affinity-clusterip-transition 80 port [tcp/http] succeeded!\n" Jun 3 11:33:50.182: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:33:50.182: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3978 exec execpod-affinityvhftz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.234.7 80' Jun 3 11:33:50.489: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.234.7 80\nConnection to 10.96.234.7 80 port [tcp/http] succeeded!\n" Jun 3 11:33:50.489: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:33:50.496: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3978 exec execpod-affinityvhftz -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.96.234.7:80/ ; done' Jun 3 11:33:50.865: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n" Jun 3 11:33:50.865: INFO: stdout: "\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt" Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:33:50.865: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:20.866: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3978 exec execpod-affinityvhftz -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.96.234.7:80/ ; done' Jun 3 11:34:21.213: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n" Jun 3 11:34:21.213: INFO: stdout: "\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-jhh84\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-45hdt\naffinity-clusterip-transition-jhh84\naffinity-clusterip-transition-jhh84" Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-jhh84 Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-45hdt Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-jhh84 Jun 3 11:34:21.213: INFO: Received response from host: affinity-clusterip-transition-jhh84 Jun 3 11:34:21.223: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-3978 exec execpod-affinityvhftz -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.96.234.7:80/ ; done' Jun 3 11:34:21.566: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.96.234.7:80/\n" Jun 3 11:34:21.567: INFO: stdout: "\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br\naffinity-clusterip-transition-ck2br" Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Received response from host: affinity-clusterip-transition-ck2br Jun 3 11:34:21.567: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-clusterip-transition in namespace services-3978, will wait for the garbage collector to delete the pods Jun 3 11:34:21.636: INFO: Deleting ReplicationController affinity-clusterip-transition took: 4.334176ms Jun 3 11:34:21.737: INFO: Terminating ReplicationController affinity-clusterip-transition pods took: 101.015717ms [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:36.247: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3978" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:60.403 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]","total":-1,"completed":47,"skipped":637,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:36.269: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating projection with secret that has name projected-secret-test-map-68acbdce-f0d1-4579-b574-e2e60fdf378b STEP: Creating a pod to test consume secrets Jun 3 11:34:36.307: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0" in namespace "projected-1919" to be "Succeeded or Failed" Jun 3 11:34:36.310: INFO: Pod "pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0": Phase="Pending", Reason="", readiness=false. Elapsed: 3.079334ms Jun 3 11:34:38.319: INFO: Pod "pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011538146s Jun 3 11:34:40.322: INFO: Pod "pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015049986s STEP: Saw pod success Jun 3 11:34:40.322: INFO: Pod "pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0" satisfied condition "Succeeded or Failed" Jun 3 11:34:40.325: INFO: Trying to get logs from node v1.21-worker pod pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0 container projected-secret-volume-test: STEP: delete the pod Jun 3 11:34:40.339: INFO: Waiting for pod pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0 to disappear Jun 3 11:34:40.342: INFO: Pod pod-projected-secrets-26c65f54-cead-4800-8b90-8d75d45677a0 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:40.342: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1919" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":48,"skipped":642,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:35.214: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name projected-configmap-test-volume-7b7c6f10-5b0d-4b13-89d2-e7bc2c2ac8ed STEP: Creating a pod to test consume configMaps Jun 3 11:34:35.252: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4" in namespace "projected-1127" to be "Succeeded or Failed" Jun 3 11:34:35.255: INFO: Pod "pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.727606ms Jun 3 11:34:37.259: INFO: Pod "pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006894085s Jun 3 11:34:39.264: INFO: Pod "pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011494458s Jun 3 11:34:41.268: INFO: Pod "pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015650526s STEP: Saw pod success Jun 3 11:34:41.268: INFO: Pod "pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4" satisfied condition "Succeeded or Failed" Jun 3 11:34:41.271: INFO: Trying to get logs from node v1.21-worker pod pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4 container agnhost-container: STEP: delete the pod Jun 3 11:34:41.284: INFO: Waiting for pod pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4 to disappear Jun 3 11:34:41.287: INFO: Pod pod-projected-configmaps-1122cd58-9e74-4cb2-a324-c8741bfad3d4 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:41.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1127" for this suite. • [SLOW TEST:6.081 seconds] [sig-storage] Projected configMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":42,"skipped":768,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:17.346: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [BeforeEach] Update Demo /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:293 [It] should scale a replication controller [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a replication controller Jun 3 11:34:17.379: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 create -f -' Jun 3 11:34:17.754: INFO: stderr: "" Jun 3 11:34:17.754: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Jun 3 11:34:17.754: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:34:17.878: INFO: stderr: "" Jun 3 11:34:17.878: INFO: stdout: "update-demo-nautilus-gfsw4 update-demo-nautilus-gksjs " Jun 3 11:34:17.878: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gfsw4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:34:17.993: INFO: stderr: "" Jun 3 11:34:17.993: INFO: stdout: "" Jun 3 11:34:17.993: INFO: update-demo-nautilus-gfsw4 is created but not running Jun 3 11:34:22.994: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:34:23.130: INFO: stderr: "" Jun 3 11:34:23.130: INFO: stdout: "update-demo-nautilus-gfsw4 update-demo-nautilus-gksjs " Jun 3 11:34:23.131: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gfsw4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:34:23.249: INFO: stderr: "" Jun 3 11:34:23.249: INFO: stdout: "true" Jun 3 11:34:23.249: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gfsw4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' Jun 3 11:34:23.355: INFO: stderr: "" Jun 3 11:34:23.355: INFO: stdout: "k8s.gcr.io/e2e-test-images/nautilus:1.4" Jun 3 11:34:23.355: INFO: validating pod update-demo-nautilus-gfsw4 Jun 3 11:34:23.359: INFO: got data: { "image": "nautilus.jpg" } Jun 3 11:34:23.359: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jun 3 11:34:23.359: INFO: update-demo-nautilus-gfsw4 is verified up and running Jun 3 11:34:23.359: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gksjs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:34:23.477: INFO: stderr: "" Jun 3 11:34:23.477: INFO: stdout: "true" Jun 3 11:34:23.477: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gksjs -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' Jun 3 11:34:23.588: INFO: stderr: "" Jun 3 11:34:23.588: INFO: stdout: "k8s.gcr.io/e2e-test-images/nautilus:1.4" Jun 3 11:34:23.588: INFO: validating pod update-demo-nautilus-gksjs Jun 3 11:34:23.592: INFO: got data: { "image": "nautilus.jpg" } Jun 3 11:34:23.592: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jun 3 11:34:23.592: INFO: update-demo-nautilus-gksjs is verified up and running STEP: scaling down the replication controller Jun 3 11:34:23.597: INFO: scanned /root for discovery docs: Jun 3 11:34:23.597: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 scale rc update-demo-nautilus --replicas=1 --timeout=5m' Jun 3 11:34:24.736: INFO: stderr: "" Jun 3 11:34:24.736: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Jun 3 11:34:24.736: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:34:24.996: INFO: stderr: "" Jun 3 11:34:24.996: INFO: stdout: "update-demo-nautilus-gfsw4 update-demo-nautilus-gksjs " STEP: Replicas for name=update-demo: expected=1 actual=2 Jun 3 11:34:29.997: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:34:30.124: INFO: stderr: "" Jun 3 11:34:30.124: INFO: stdout: "update-demo-nautilus-gfsw4 update-demo-nautilus-gksjs " STEP: Replicas for name=update-demo: expected=1 actual=2 Jun 3 11:34:35.125: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:34:35.244: INFO: stderr: "" Jun 3 11:34:35.244: INFO: stdout: "update-demo-nautilus-gksjs " Jun 3 11:34:35.244: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gksjs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:34:35.365: INFO: stderr: "" Jun 3 11:34:35.365: INFO: stdout: "true" Jun 3 11:34:35.365: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gksjs -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' Jun 3 11:34:35.482: INFO: stderr: "" Jun 3 11:34:35.482: INFO: stdout: "k8s.gcr.io/e2e-test-images/nautilus:1.4" Jun 3 11:34:35.482: INFO: validating pod update-demo-nautilus-gksjs Jun 3 11:34:35.485: INFO: got data: { "image": "nautilus.jpg" } Jun 3 11:34:35.485: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jun 3 11:34:35.485: INFO: update-demo-nautilus-gksjs is verified up and running STEP: scaling up the replication controller Jun 3 11:34:35.490: INFO: scanned /root for discovery docs: Jun 3 11:34:35.490: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 scale rc update-demo-nautilus --replicas=2 --timeout=5m' Jun 3 11:34:36.630: INFO: stderr: "" Jun 3 11:34:36.630: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Jun 3 11:34:36.630: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:34:36.755: INFO: stderr: "" Jun 3 11:34:36.755: INFO: stdout: "update-demo-nautilus-dx8s4 update-demo-nautilus-gksjs " Jun 3 11:34:36.755: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-dx8s4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:34:36.873: INFO: stderr: "" Jun 3 11:34:36.873: INFO: stdout: "" Jun 3 11:34:36.873: INFO: update-demo-nautilus-dx8s4 is created but not running Jun 3 11:34:41.873: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' Jun 3 11:34:42.002: INFO: stderr: "" Jun 3 11:34:42.002: INFO: stdout: "update-demo-nautilus-dx8s4 update-demo-nautilus-gksjs " Jun 3 11:34:42.002: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-dx8s4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:34:42.121: INFO: stderr: "" Jun 3 11:34:42.121: INFO: stdout: "true" Jun 3 11:34:42.121: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-dx8s4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' Jun 3 11:34:42.223: INFO: stderr: "" Jun 3 11:34:42.223: INFO: stdout: "k8s.gcr.io/e2e-test-images/nautilus:1.4" Jun 3 11:34:42.223: INFO: validating pod update-demo-nautilus-dx8s4 Jun 3 11:34:42.228: INFO: got data: { "image": "nautilus.jpg" } Jun 3 11:34:42.228: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jun 3 11:34:42.228: INFO: update-demo-nautilus-dx8s4 is verified up and running Jun 3 11:34:42.228: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gksjs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' Jun 3 11:34:42.343: INFO: stderr: "" Jun 3 11:34:42.343: INFO: stdout: "true" Jun 3 11:34:42.343: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods update-demo-nautilus-gksjs -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' Jun 3 11:34:42.438: INFO: stderr: "" Jun 3 11:34:42.438: INFO: stdout: "k8s.gcr.io/e2e-test-images/nautilus:1.4" Jun 3 11:34:42.438: INFO: validating pod update-demo-nautilus-gksjs Jun 3 11:34:42.441: INFO: got data: { "image": "nautilus.jpg" } Jun 3 11:34:42.441: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jun 3 11:34:42.441: INFO: update-demo-nautilus-gksjs is verified up and running STEP: using delete to clean up resources Jun 3 11:34:42.441: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 delete --grace-period=0 --force -f -' Jun 3 11:34:42.552: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jun 3 11:34:42.552: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Jun 3 11:34:42.552: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get rc,svc -l name=update-demo --no-headers' Jun 3 11:34:42.670: INFO: stderr: "No resources found in kubectl-7306 namespace.\n" Jun 3 11:34:42.670: INFO: stdout: "" Jun 3 11:34:42.670: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7306 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Jun 3 11:34:42.785: INFO: stderr: "" Jun 3 11:34:42.785: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:42.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7306" for this suite. • [SLOW TEST:25.450 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Update Demo /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:291 should scale a replication controller [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance]","total":-1,"completed":18,"skipped":278,"failed":0} SSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:41.342: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:69 [It] should update/patch PodDisruptionBudget status [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Waiting for the pdb to be processed STEP: Updating PodDisruptionBudget status STEP: Waiting for all pods to be running Jun 3 11:34:43.388: INFO: running pods: 0 < 1 STEP: locating a running pod STEP: Waiting for the pdb to be processed STEP: Patching PodDisruptionBudget status STEP: Waiting for the pdb to be processed [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:45.420: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-3241" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController should update/patch PodDisruptionBudget status [Conformance]","total":-1,"completed":43,"skipped":792,"failed":0} SSSSSSSSSSS ------------------------------ [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:42.019: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746 [It] should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating service in namespace services-4223 Jun 3 11:33:42.053: INFO: The status of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:44.057: INFO: The status of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:46.057: INFO: The status of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:33:48.058: INFO: The status of Pod kube-proxy-mode-detector is Running (Ready = true) Jun 3 11:33:48.061: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' Jun 3 11:33:48.309: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" Jun 3 11:33:48.309: INFO: stdout: "iptables" Jun 3 11:33:48.309: INFO: proxyMode: iptables Jun 3 11:33:48.315: INFO: Waiting for pod kube-proxy-mode-detector to disappear Jun 3 11:33:48.319: INFO: Pod kube-proxy-mode-detector no longer exists STEP: creating service affinity-nodeport-timeout in namespace services-4223 STEP: creating replication controller affinity-nodeport-timeout in namespace services-4223 I0603 11:33:48.332580 19 runners.go:190] Created replication controller with name: affinity-nodeport-timeout, namespace: services-4223, replica count: 3 I0603 11:33:51.384224 19 runners.go:190] affinity-nodeport-timeout Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:33:54.385453 19 runners.go:190] affinity-nodeport-timeout Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:33:57.386564 19 runners.go:190] affinity-nodeport-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:33:57.479: INFO: Creating new exec pod Jun 3 11:34:06.495: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec execpod-affinitybbzst -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-timeout 80' Jun 3 11:34:06.716: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-timeout 80\nConnection to affinity-nodeport-timeout 80 port [tcp/http] succeeded!\n" Jun 3 11:34:06.716: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:34:06.716: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec execpod-affinitybbzst -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.96.235.242 80' Jun 3 11:34:06.932: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.96.235.242 80\nConnection to 10.96.235.242 80 port [tcp/http] succeeded!\n" Jun 3 11:34:06.932: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:34:06.932: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec execpod-affinitybbzst -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.4 32476' Jun 3 11:34:07.229: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.4 32476\nConnection to 172.18.0.4 32476 port [tcp/*] succeeded!\n" Jun 3 11:34:07.229: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:34:07.229: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec execpod-affinitybbzst -- /bin/sh -x -c echo hostName | nc -v -t -w 2 172.18.0.2 32476' Jun 3 11:34:07.490: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 172.18.0.2 32476\nConnection to 172.18.0.2 32476 port [tcp/*] succeeded!\n" Jun 3 11:34:07.490: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" Jun 3 11:34:07.490: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec execpod-affinitybbzst -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.4:32476/ ; done' Jun 3 11:34:07.856: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n" Jun 3 11:34:07.856: INFO: stdout: "\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp\naffinity-nodeport-timeout-j8jxp" Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.856: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.857: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.857: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.857: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.857: INFO: Received response from host: affinity-nodeport-timeout-j8jxp Jun 3 11:34:07.857: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec execpod-affinitybbzst -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://172.18.0.4:32476/' Jun 3 11:34:08.105: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n" Jun 3 11:34:08.105: INFO: stdout: "affinity-nodeport-timeout-j8jxp" Jun 3 11:34:28.107: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=services-4223 exec execpod-affinitybbzst -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://172.18.0.4:32476/' Jun 3 11:34:28.376: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://172.18.0.4:32476/\n" Jun 3 11:34:28.376: INFO: stdout: "affinity-nodeport-timeout-78r6r" Jun 3 11:34:28.376: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-nodeport-timeout in namespace services-4223, will wait for the garbage collector to delete the pods Jun 3 11:34:28.443: INFO: Deleting ReplicationController affinity-nodeport-timeout took: 4.604964ms Jun 3 11:34:28.544: INFO: Terminating ReplicationController affinity-nodeport-timeout pods took: 101.113153ms [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:47.258: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-4223" for this suite. [AfterEach] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750 • [SLOW TEST:65.246 seconds] [sig-network] Services /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance]","total":-1,"completed":34,"skipped":539,"failed":0} SS ------------------------------ [BeforeEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:47.271: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: creating a watch on configmaps STEP: creating a new configmap STEP: modifying the configmap once STEP: closing the watch once it receives two notifications Jun 3 11:34:47.315: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-2800 9cefba97-20d9-4689-8296-23d12762f716 3299934 0 2021-06-03 11:34:47 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2021-06-03 11:34:47 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:34:47.315: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-2800 9cefba97-20d9-4689-8296-23d12762f716 3299935 0 2021-06-03 11:34:47 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2021-06-03 11:34:47 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying the configmap a second time, while the watch is closed STEP: creating a new watch on configmaps from the last resource version observed by the first watch STEP: deleting the configmap STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed Jun 3 11:34:47.326: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-2800 9cefba97-20d9-4689-8296-23d12762f716 3299936 0 2021-06-03 11:34:47 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2021-06-03 11:34:47 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Jun 3 11:34:47.327: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-2800 9cefba97-20d9-4689-8296-23d12762f716 3299937 0 2021-06-03 11:34:47 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2021-06-03 11:34:47 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:47.327: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-2800" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":-1,"completed":35,"skipped":541,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:40.391: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should fail substituting values in a volume subpath with backticks [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:34:42.437: INFO: Deleting pod "var-expansion-61a239cb-a843-4279-a6ef-96614ad33ab6" in namespace "var-expansion-2343" Jun 3 11:34:42.441: INFO: Wait up to 5m0s for pod "var-expansion-61a239cb-a843-4279-a6ef-96614ad33ab6" to be fully deleted [AfterEach] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:50.448: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-2343" for this suite. • [SLOW TEST:10.065 seconds] [sig-node] Variable Expansion /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should fail substituting values in a volume subpath with backticks [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Variable Expansion should fail substituting values in a volume subpath with backticks [Slow] [Conformance]","total":-1,"completed":49,"skipped":663,"failed":0} SS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:11.073: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of same group but different versions [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation Jun 3 11:34:11.112: INFO: >>> kubeConfig: /root/.kube/config STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation Jun 3 11:34:28.749: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:34:33.333: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:51.585: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-5324" for this suite. • [SLOW TEST:40.522 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of same group but different versions [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":-1,"completed":22,"skipped":487,"failed":0} SSSSSS ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","total":-1,"completed":44,"skipped":717,"failed":0} [BeforeEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:49.166: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for deployment deletion to see if the garbage collector mistakenly deletes the rs STEP: Gathering metrics W0603 11:33:50.250750 34 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Jun 3 11:34:52.269: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:52.269: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-8999" for this suite. • [SLOW TEST:63.114 seconds] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","total":-1,"completed":45,"skipped":717,"failed":0} SSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:50.461: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should provide container's memory request [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:34:50.496: INFO: Waiting up to 5m0s for pod "downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54" in namespace "projected-6736" to be "Succeeded or Failed" Jun 3 11:34:50.498: INFO: Pod "downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00999ms Jun 3 11:34:52.505: INFO: Pod "downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00908987s Jun 3 11:34:54.509: INFO: Pod "downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013777941s STEP: Saw pod success Jun 3 11:34:54.510: INFO: Pod "downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54" satisfied condition "Succeeded or Failed" Jun 3 11:34:54.513: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54 container client-container: STEP: delete the pod Jun 3 11:34:54.530: INFO: Waiting for pod downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54 to disappear Jun 3 11:34:54.534: INFO: Pod downwardapi-volume-624de54d-38a4-44b2-bc36-42ee450c4d54 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:54.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6736" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":-1,"completed":50,"skipped":665,"failed":0} SS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:51.607: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:34:51.642: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8" in namespace "projected-1543" to be "Succeeded or Failed" Jun 3 11:34:51.645: INFO: Pod "downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.604479ms Jun 3 11:34:53.649: INFO: Pod "downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006829433s Jun 3 11:34:55.653: INFO: Pod "downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010811659s STEP: Saw pod success Jun 3 11:34:55.653: INFO: Pod "downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8" satisfied condition "Succeeded or Failed" Jun 3 11:34:55.656: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8 container client-container: STEP: delete the pod Jun 3 11:34:55.670: INFO: Waiting for pod downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8 to disappear Jun 3 11:34:55.674: INFO: Pod downwardapi-volume-0290e804-8b4a-4f84-9163-6c89d29543e8 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:55.674: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1543" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":-1,"completed":23,"skipped":493,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-node] Lease /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:55.701: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename lease-test STEP: Waiting for a default service account to be provisioned in namespace [It] lease API should be available [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 [AfterEach] [sig-node] Lease /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:55.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "lease-test-7003" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Lease lease API should be available [Conformance]","total":-1,"completed":24,"skipped":500,"failed":0} SS ------------------------------ [BeforeEach] [sig-network] Service endpoints latency /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:47.360: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svc-latency STEP: Waiting for a default service account to be provisioned in namespace [It] should not be very high [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:34:47.385: INFO: >>> kubeConfig: /root/.kube/config STEP: creating replication controller svc-latency-rc in namespace svc-latency-2880 I0603 11:34:47.402251 19 runners.go:190] Created replication controller with name: svc-latency-rc, namespace: svc-latency-2880, replica count: 1 I0603 11:34:48.454226 19 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:34:49.454371 19 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0603 11:34:50.455326 19 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jun 3 11:34:50.563: INFO: Created: latency-svc-x2qxw Jun 3 11:34:50.567: INFO: Got endpoints: latency-svc-x2qxw [12.303063ms] Jun 3 11:34:50.574: INFO: Created: latency-svc-zqwjv Jun 3 11:34:50.576: INFO: Created: latency-svc-vtj9r Jun 3 11:34:50.577: INFO: Got endpoints: latency-svc-zqwjv [9.87542ms] Jun 3 11:34:50.579: INFO: Created: latency-svc-sv5d9 Jun 3 11:34:50.581: INFO: Got endpoints: latency-svc-vtj9r [13.238305ms] Jun 3 11:34:50.581: INFO: Got endpoints: latency-svc-sv5d9 [13.788118ms] Jun 3 11:34:50.583: INFO: Created: latency-svc-sppb7 Jun 3 11:34:50.585: INFO: Created: latency-svc-l9nc7 Jun 3 11:34:50.586: INFO: Got endpoints: latency-svc-sppb7 [17.976461ms] Jun 3 11:34:50.588: INFO: Created: latency-svc-fj4lm Jun 3 11:34:50.588: INFO: Got endpoints: latency-svc-l9nc7 [20.026816ms] Jun 3 11:34:50.590: INFO: Created: latency-svc-bwfm8 Jun 3 11:34:50.591: INFO: Got endpoints: latency-svc-fj4lm [22.519319ms] Jun 3 11:34:50.593: INFO: Created: latency-svc-d8xq5 Jun 3 11:34:50.593: INFO: Got endpoints: latency-svc-bwfm8 [25.303108ms] Jun 3 11:34:50.595: INFO: Created: latency-svc-lphrw Jun 3 11:34:50.596: INFO: Got endpoints: latency-svc-d8xq5 [28.068618ms] Jun 3 11:34:50.598: INFO: Got endpoints: latency-svc-lphrw [29.609268ms] Jun 3 11:34:50.598: INFO: Created: latency-svc-827s2 Jun 3 11:34:50.600: INFO: Created: latency-svc-48cbv Jun 3 11:34:50.601: INFO: Got endpoints: latency-svc-827s2 [32.617443ms] Jun 3 11:34:50.602: INFO: Created: latency-svc-478k9 Jun 3 11:34:50.602: INFO: Got endpoints: latency-svc-48cbv [34.552025ms] Jun 3 11:34:50.604: INFO: Created: latency-svc-l4dj9 Jun 3 11:34:50.604: INFO: Got endpoints: latency-svc-478k9 [36.464436ms] Jun 3 11:34:50.606: INFO: Created: latency-svc-kncxt Jun 3 11:34:50.607: INFO: Got endpoints: latency-svc-l4dj9 [39.095456ms] Jun 3 11:34:50.608: INFO: Got endpoints: latency-svc-kncxt [40.022738ms] Jun 3 11:34:50.608: INFO: Created: latency-svc-m5vlx Jun 3 11:34:50.611: INFO: Created: latency-svc-jkrwk Jun 3 11:34:50.611: INFO: Got endpoints: latency-svc-m5vlx [43.131053ms] Jun 3 11:34:50.613: INFO: Created: latency-svc-5j8x4 Jun 3 11:34:50.614: INFO: Got endpoints: latency-svc-jkrwk [36.078839ms] Jun 3 11:34:50.615: INFO: Created: latency-svc-z7ctj Jun 3 11:34:50.616: INFO: Got endpoints: latency-svc-5j8x4 [34.813471ms] Jun 3 11:34:50.617: INFO: Created: latency-svc-gxnz6 Jun 3 11:34:50.618: INFO: Got endpoints: latency-svc-z7ctj [36.049984ms] Jun 3 11:34:50.620: INFO: Created: latency-svc-hrhcn Jun 3 11:34:50.620: INFO: Got endpoints: latency-svc-gxnz6 [34.228571ms] Jun 3 11:34:50.622: INFO: Created: latency-svc-rj2wt Jun 3 11:34:50.622: INFO: Got endpoints: latency-svc-hrhcn [34.367523ms] Jun 3 11:34:50.624: INFO: Created: latency-svc-xkbhl Jun 3 11:34:50.625: INFO: Got endpoints: latency-svc-rj2wt [34.038152ms] Jun 3 11:34:50.626: INFO: Created: latency-svc-zcfdq Jun 3 11:34:50.626: INFO: Got endpoints: latency-svc-xkbhl [32.77836ms] Jun 3 11:34:50.629: INFO: Got endpoints: latency-svc-zcfdq [32.715961ms] Jun 3 11:34:50.629: INFO: Created: latency-svc-6tc2k Jun 3 11:34:50.630: INFO: Created: latency-svc-8sd9m Jun 3 11:34:50.632: INFO: Got endpoints: latency-svc-6tc2k [34.184118ms] Jun 3 11:34:50.632: INFO: Created: latency-svc-zk94w Jun 3 11:34:50.633: INFO: Got endpoints: latency-svc-8sd9m [32.539536ms] Jun 3 11:34:50.634: INFO: Created: latency-svc-mbdvq Jun 3 11:34:50.634: INFO: Got endpoints: latency-svc-zk94w [31.997717ms] Jun 3 11:34:50.636: INFO: Created: latency-svc-hz9mb Jun 3 11:34:50.637: INFO: Got endpoints: latency-svc-mbdvq [32.289834ms] Jun 3 11:34:50.638: INFO: Created: latency-svc-sfnz8 Jun 3 11:34:50.639: INFO: Got endpoints: latency-svc-hz9mb [32.091352ms] Jun 3 11:34:50.640: INFO: Got endpoints: latency-svc-sfnz8 [32.521209ms] Jun 3 11:34:50.641: INFO: Created: latency-svc-wkj7d Jun 3 11:34:50.643: INFO: Created: latency-svc-hqps2 Jun 3 11:34:50.644: INFO: Got endpoints: latency-svc-wkj7d [32.977599ms] Jun 3 11:34:50.645: INFO: Created: latency-svc-9lzq6 Jun 3 11:34:50.647: INFO: Created: latency-svc-chtlq Jun 3 11:34:50.649: INFO: Created: latency-svc-zsnfl Jun 3 11:34:50.658: INFO: Created: latency-svc-rgc52 Jun 3 11:34:50.659: INFO: Created: latency-svc-tgc5b Jun 3 11:34:50.662: INFO: Created: latency-svc-cdzkv Jun 3 11:34:50.664: INFO: Created: latency-svc-2kg48 Jun 3 11:34:50.666: INFO: Created: latency-svc-4sxnt Jun 3 11:34:50.666: INFO: Got endpoints: latency-svc-hqps2 [52.583313ms] Jun 3 11:34:50.668: INFO: Created: latency-svc-vk9j2 Jun 3 11:34:50.670: INFO: Created: latency-svc-lctlg Jun 3 11:34:50.672: INFO: Created: latency-svc-lrhkj Jun 3 11:34:50.674: INFO: Created: latency-svc-kg5vl Jun 3 11:34:50.677: INFO: Created: latency-svc-bq4jg Jun 3 11:34:50.679: INFO: Created: latency-svc-xlrm9 Jun 3 11:34:50.684: INFO: Created: latency-svc-d4c6r Jun 3 11:34:50.718: INFO: Got endpoints: latency-svc-9lzq6 [101.730529ms] Jun 3 11:34:50.724: INFO: Created: latency-svc-gzh7x Jun 3 11:34:50.768: INFO: Got endpoints: latency-svc-chtlq [150.059678ms] Jun 3 11:34:50.775: INFO: Created: latency-svc-mbc2g Jun 3 11:34:50.818: INFO: Got endpoints: latency-svc-zsnfl [197.625536ms] Jun 3 11:34:50.825: INFO: Created: latency-svc-cjgc7 Jun 3 11:34:50.871: INFO: Got endpoints: latency-svc-rgc52 [248.495338ms] Jun 3 11:34:50.879: INFO: Created: latency-svc-blb4s Jun 3 11:34:50.918: INFO: Got endpoints: latency-svc-tgc5b [293.371425ms] Jun 3 11:34:50.926: INFO: Created: latency-svc-69ckp Jun 3 11:34:50.967: INFO: Got endpoints: latency-svc-cdzkv [340.664561ms] Jun 3 11:34:50.974: INFO: Created: latency-svc-bgzx9 Jun 3 11:34:51.018: INFO: Got endpoints: latency-svc-2kg48 [389.27356ms] Jun 3 11:34:51.025: INFO: Created: latency-svc-zd2zd Jun 3 11:34:51.068: INFO: Got endpoints: latency-svc-4sxnt [435.933839ms] Jun 3 11:34:51.075: INFO: Created: latency-svc-b8cpb Jun 3 11:34:51.118: INFO: Got endpoints: latency-svc-vk9j2 [484.481087ms] Jun 3 11:34:51.124: INFO: Created: latency-svc-n97s7 Jun 3 11:34:51.168: INFO: Got endpoints: latency-svc-lctlg [533.558889ms] Jun 3 11:34:51.176: INFO: Created: latency-svc-lcw2b Jun 3 11:34:51.217: INFO: Got endpoints: latency-svc-lrhkj [580.393101ms] Jun 3 11:34:51.224: INFO: Created: latency-svc-t5xv8 Jun 3 11:34:51.267: INFO: Got endpoints: latency-svc-kg5vl [627.903078ms] Jun 3 11:34:51.274: INFO: Created: latency-svc-lflhd Jun 3 11:34:51.317: INFO: Got endpoints: latency-svc-bq4jg [676.846872ms] Jun 3 11:34:51.325: INFO: Created: latency-svc-wfcpt Jun 3 11:34:51.373: INFO: Got endpoints: latency-svc-xlrm9 [728.567243ms] Jun 3 11:34:51.380: INFO: Created: latency-svc-hz4b5 Jun 3 11:34:51.418: INFO: Got endpoints: latency-svc-d4c6r [751.318902ms] Jun 3 11:34:51.424: INFO: Created: latency-svc-fhlbd Jun 3 11:34:51.467: INFO: Got endpoints: latency-svc-gzh7x [749.197824ms] Jun 3 11:34:51.474: INFO: Created: latency-svc-98c6b Jun 3 11:34:51.578: INFO: Got endpoints: latency-svc-mbc2g [809.864941ms] Jun 3 11:34:51.578: INFO: Got endpoints: latency-svc-cjgc7 [759.961986ms] Jun 3 11:34:51.586: INFO: Created: latency-svc-c4mm4 Jun 3 11:34:51.590: INFO: Created: latency-svc-6xd5p Jun 3 11:34:51.618: INFO: Got endpoints: latency-svc-blb4s [746.489147ms] Jun 3 11:34:51.624: INFO: Created: latency-svc-nc8fk Jun 3 11:34:51.667: INFO: Got endpoints: latency-svc-69ckp [748.977167ms] Jun 3 11:34:51.676: INFO: Created: latency-svc-6xtsw Jun 3 11:34:51.718: INFO: Got endpoints: latency-svc-bgzx9 [750.812216ms] Jun 3 11:34:51.725: INFO: Created: latency-svc-bh6hw Jun 3 11:34:51.768: INFO: Got endpoints: latency-svc-zd2zd [749.8008ms] Jun 3 11:34:51.775: INFO: Created: latency-svc-fj5wp Jun 3 11:34:51.817: INFO: Got endpoints: latency-svc-b8cpb [748.6822ms] Jun 3 11:34:51.824: INFO: Created: latency-svc-r6j98 Jun 3 11:34:51.868: INFO: Got endpoints: latency-svc-n97s7 [750.23313ms] Jun 3 11:34:51.876: INFO: Created: latency-svc-cpjtn Jun 3 11:34:51.917: INFO: Got endpoints: latency-svc-lcw2b [749.307528ms] Jun 3 11:34:51.925: INFO: Created: latency-svc-hqgbf Jun 3 11:34:51.967: INFO: Got endpoints: latency-svc-t5xv8 [749.44224ms] Jun 3 11:34:51.974: INFO: Created: latency-svc-45f5v Jun 3 11:34:52.018: INFO: Got endpoints: latency-svc-lflhd [751.171216ms] Jun 3 11:34:52.025: INFO: Created: latency-svc-v6bdm Jun 3 11:34:52.067: INFO: Got endpoints: latency-svc-wfcpt [749.318685ms] Jun 3 11:34:52.075: INFO: Created: latency-svc-bwffg Jun 3 11:34:52.118: INFO: Got endpoints: latency-svc-hz4b5 [745.17734ms] Jun 3 11:34:52.126: INFO: Created: latency-svc-vkvr2 Jun 3 11:34:52.168: INFO: Got endpoints: latency-svc-fhlbd [750.589726ms] Jun 3 11:34:52.176: INFO: Created: latency-svc-76clh Jun 3 11:34:52.217: INFO: Got endpoints: latency-svc-98c6b [749.957592ms] Jun 3 11:34:52.224: INFO: Created: latency-svc-kcddj Jun 3 11:34:52.267: INFO: Got endpoints: latency-svc-c4mm4 [689.044239ms] Jun 3 11:34:52.274: INFO: Created: latency-svc-rbwhr Jun 3 11:34:52.318: INFO: Got endpoints: latency-svc-6xd5p [739.928326ms] Jun 3 11:34:52.325: INFO: Created: latency-svc-h7hjg Jun 3 11:34:52.367: INFO: Got endpoints: latency-svc-nc8fk [749.438154ms] Jun 3 11:34:52.374: INFO: Created: latency-svc-824hv Jun 3 11:34:52.416: INFO: Got endpoints: latency-svc-6xtsw [748.79845ms] Jun 3 11:34:52.422: INFO: Created: latency-svc-jnqmv Jun 3 11:34:52.466: INFO: Got endpoints: latency-svc-bh6hw [748.32228ms] Jun 3 11:34:52.473: INFO: Created: latency-svc-xhr7p Jun 3 11:34:52.516: INFO: Got endpoints: latency-svc-fj5wp [748.396963ms] Jun 3 11:34:52.521: INFO: Created: latency-svc-c2wxs Jun 3 11:34:52.567: INFO: Got endpoints: latency-svc-r6j98 [750.697822ms] Jun 3 11:34:52.574: INFO: Created: latency-svc-nfrpz Jun 3 11:34:52.618: INFO: Got endpoints: latency-svc-cpjtn [749.491221ms] Jun 3 11:34:52.624: INFO: Created: latency-svc-nj42z Jun 3 11:34:52.667: INFO: Got endpoints: latency-svc-hqgbf [749.51534ms] Jun 3 11:34:52.674: INFO: Created: latency-svc-gdg6j Jun 3 11:34:52.718: INFO: Got endpoints: latency-svc-45f5v [750.698794ms] Jun 3 11:34:52.725: INFO: Created: latency-svc-qrsjj Jun 3 11:34:52.770: INFO: Got endpoints: latency-svc-v6bdm [751.387936ms] Jun 3 11:34:52.777: INFO: Created: latency-svc-s8bxm Jun 3 11:34:52.818: INFO: Got endpoints: latency-svc-bwffg [750.750839ms] Jun 3 11:34:52.825: INFO: Created: latency-svc-7nrz5 Jun 3 11:34:52.867: INFO: Got endpoints: latency-svc-vkvr2 [748.360345ms] Jun 3 11:34:52.874: INFO: Created: latency-svc-t6dbj Jun 3 11:34:52.918: INFO: Got endpoints: latency-svc-76clh [749.404883ms] Jun 3 11:34:52.925: INFO: Created: latency-svc-bk92h Jun 3 11:34:52.968: INFO: Got endpoints: latency-svc-kcddj [751.345041ms] Jun 3 11:34:52.975: INFO: Created: latency-svc-mmn99 Jun 3 11:34:53.017: INFO: Got endpoints: latency-svc-rbwhr [750.069655ms] Jun 3 11:34:53.024: INFO: Created: latency-svc-v8xt4 Jun 3 11:34:53.068: INFO: Got endpoints: latency-svc-h7hjg [749.91777ms] Jun 3 11:34:53.075: INFO: Created: latency-svc-xbbmf Jun 3 11:34:53.119: INFO: Got endpoints: latency-svc-824hv [751.248869ms] Jun 3 11:34:53.126: INFO: Created: latency-svc-6hj54 Jun 3 11:34:53.167: INFO: Got endpoints: latency-svc-jnqmv [751.146225ms] Jun 3 11:34:53.175: INFO: Created: latency-svc-n6rmf Jun 3 11:34:53.217: INFO: Got endpoints: latency-svc-xhr7p [750.835481ms] Jun 3 11:34:53.225: INFO: Created: latency-svc-zsbk5 Jun 3 11:34:53.267: INFO: Got endpoints: latency-svc-c2wxs [751.179455ms] Jun 3 11:34:53.274: INFO: Created: latency-svc-scv9x Jun 3 11:34:53.317: INFO: Got endpoints: latency-svc-nfrpz [749.835173ms] Jun 3 11:34:53.330: INFO: Created: latency-svc-t7xrv Jun 3 11:34:53.367: INFO: Got endpoints: latency-svc-nj42z [749.778069ms] Jun 3 11:34:53.375: INFO: Created: latency-svc-92ct9 Jun 3 11:34:53.417: INFO: Got endpoints: latency-svc-gdg6j [750.343748ms] Jun 3 11:34:53.427: INFO: Created: latency-svc-qzhm6 Jun 3 11:34:53.468: INFO: Got endpoints: latency-svc-qrsjj [749.992355ms] Jun 3 11:34:53.476: INFO: Created: latency-svc-7f27q Jun 3 11:34:53.518: INFO: Got endpoints: latency-svc-s8bxm [747.810053ms] Jun 3 11:34:53.526: INFO: Created: latency-svc-x5c2c Jun 3 11:34:53.568: INFO: Got endpoints: latency-svc-7nrz5 [750.205281ms] Jun 3 11:34:53.576: INFO: Created: latency-svc-bhlst Jun 3 11:34:53.618: INFO: Got endpoints: latency-svc-t6dbj [751.180699ms] Jun 3 11:34:53.625: INFO: Created: latency-svc-7dfbr Jun 3 11:34:53.668: INFO: Got endpoints: latency-svc-bk92h [750.338938ms] Jun 3 11:34:53.676: INFO: Created: latency-svc-wncvf Jun 3 11:34:53.718: INFO: Got endpoints: latency-svc-mmn99 [749.914135ms] Jun 3 11:34:53.725: INFO: Created: latency-svc-llrz5 Jun 3 11:34:53.768: INFO: Got endpoints: latency-svc-v8xt4 [750.779467ms] Jun 3 11:34:53.775: INFO: Created: latency-svc-lb728 Jun 3 11:34:53.817: INFO: Got endpoints: latency-svc-xbbmf [748.70382ms] Jun 3 11:34:53.824: INFO: Created: latency-svc-xp8gc Jun 3 11:34:53.868: INFO: Got endpoints: latency-svc-6hj54 [749.028459ms] Jun 3 11:34:53.875: INFO: Created: latency-svc-dz6cb Jun 3 11:34:53.918: INFO: Got endpoints: latency-svc-n6rmf [750.798006ms] Jun 3 11:34:53.925: INFO: Created: latency-svc-rrnms Jun 3 11:34:53.968: INFO: Got endpoints: latency-svc-zsbk5 [750.542322ms] Jun 3 11:34:53.975: INFO: Created: latency-svc-9gz75 Jun 3 11:34:54.018: INFO: Got endpoints: latency-svc-scv9x [750.191213ms] Jun 3 11:34:54.025: INFO: Created: latency-svc-4wvv4 Jun 3 11:34:54.068: INFO: Got endpoints: latency-svc-t7xrv [751.231235ms] Jun 3 11:34:54.075: INFO: Created: latency-svc-qlt7w Jun 3 11:34:54.119: INFO: Got endpoints: latency-svc-92ct9 [751.555077ms] Jun 3 11:34:54.127: INFO: Created: latency-svc-jjn2q Jun 3 11:34:54.169: INFO: Got endpoints: latency-svc-qzhm6 [751.210854ms] Jun 3 11:34:54.177: INFO: Created: latency-svc-bmgvm Jun 3 11:34:54.217: INFO: Got endpoints: latency-svc-7f27q [749.222213ms] Jun 3 11:34:54.224: INFO: Created: latency-svc-bpp2x Jun 3 11:34:54.268: INFO: Got endpoints: latency-svc-x5c2c [750.425648ms] Jun 3 11:34:54.276: INFO: Created: latency-svc-lrqdt Jun 3 11:34:54.317: INFO: Got endpoints: latency-svc-bhlst [749.176223ms] Jun 3 11:34:54.325: INFO: Created: latency-svc-75ndm Jun 3 11:34:54.369: INFO: Got endpoints: latency-svc-7dfbr [751.546969ms] Jun 3 11:34:54.376: INFO: Created: latency-svc-c85nz Jun 3 11:34:54.417: INFO: Got endpoints: latency-svc-wncvf [748.797074ms] Jun 3 11:34:54.424: INFO: Created: latency-svc-2ms6z Jun 3 11:34:54.468: INFO: Got endpoints: latency-svc-llrz5 [749.259839ms] Jun 3 11:34:54.475: INFO: Created: latency-svc-7qnjp Jun 3 11:34:54.519: INFO: Got endpoints: latency-svc-lb728 [750.525393ms] Jun 3 11:34:54.525: INFO: Created: latency-svc-pb9sp Jun 3 11:34:54.567: INFO: Got endpoints: latency-svc-xp8gc [750.25726ms] Jun 3 11:34:54.575: INFO: Created: latency-svc-dxzg7 Jun 3 11:34:54.617: INFO: Got endpoints: latency-svc-dz6cb [749.571192ms] Jun 3 11:34:54.625: INFO: Created: latency-svc-trmnt Jun 3 11:34:54.671: INFO: Got endpoints: latency-svc-rrnms [752.971955ms] Jun 3 11:34:54.682: INFO: Created: latency-svc-qn6c4 Jun 3 11:34:54.717: INFO: Got endpoints: latency-svc-9gz75 [749.065772ms] Jun 3 11:34:54.724: INFO: Created: latency-svc-cz6z2 Jun 3 11:34:54.768: INFO: Got endpoints: latency-svc-4wvv4 [750.175987ms] Jun 3 11:34:54.775: INFO: Created: latency-svc-4tzhw Jun 3 11:34:54.818: INFO: Got endpoints: latency-svc-qlt7w [748.985013ms] Jun 3 11:34:54.825: INFO: Created: latency-svc-kzp79 Jun 3 11:34:54.869: INFO: Got endpoints: latency-svc-jjn2q [749.575615ms] Jun 3 11:34:54.876: INFO: Created: latency-svc-bvt48 Jun 3 11:34:54.918: INFO: Got endpoints: latency-svc-bmgvm [749.077887ms] Jun 3 11:34:54.925: INFO: Created: latency-svc-5m788 Jun 3 11:34:54.968: INFO: Got endpoints: latency-svc-bpp2x [750.597367ms] Jun 3 11:34:54.975: INFO: Created: latency-svc-nq8sn Jun 3 11:34:55.017: INFO: Got endpoints: latency-svc-lrqdt [749.217395ms] Jun 3 11:34:55.024: INFO: Created: latency-svc-btvv5 Jun 3 11:34:55.067: INFO: Got endpoints: latency-svc-75ndm [749.941698ms] Jun 3 11:34:55.074: INFO: Created: latency-svc-wq4fv Jun 3 11:34:55.117: INFO: Got endpoints: latency-svc-c85nz [747.374382ms] Jun 3 11:34:55.124: INFO: Created: latency-svc-mlrrh Jun 3 11:34:55.167: INFO: Got endpoints: latency-svc-2ms6z [750.190436ms] Jun 3 11:34:55.174: INFO: Created: latency-svc-8rtzg Jun 3 11:34:55.216: INFO: Got endpoints: latency-svc-7qnjp [748.789414ms] Jun 3 11:34:55.223: INFO: Created: latency-svc-hmp8p Jun 3 11:34:55.265: INFO: Got endpoints: latency-svc-pb9sp [746.802593ms] Jun 3 11:34:55.272: INFO: Created: latency-svc-pzqpg Jun 3 11:34:55.318: INFO: Got endpoints: latency-svc-dxzg7 [751.308553ms] Jun 3 11:34:55.326: INFO: Created: latency-svc-28rpf Jun 3 11:34:55.369: INFO: Got endpoints: latency-svc-trmnt [751.440851ms] Jun 3 11:34:55.376: INFO: Created: latency-svc-bc6h7 Jun 3 11:34:55.417: INFO: Got endpoints: latency-svc-qn6c4 [746.019982ms] Jun 3 11:34:55.425: INFO: Created: latency-svc-frkxr Jun 3 11:34:55.468: INFO: Got endpoints: latency-svc-cz6z2 [751.474263ms] Jun 3 11:34:55.475: INFO: Created: latency-svc-w4t4x Jun 3 11:34:55.517: INFO: Got endpoints: latency-svc-4tzhw [749.185752ms] Jun 3 11:34:55.524: INFO: Created: latency-svc-kl4gh Jun 3 11:34:55.567: INFO: Got endpoints: latency-svc-kzp79 [749.201704ms] Jun 3 11:34:55.574: INFO: Created: latency-svc-rpwv8 Jun 3 11:34:55.617: INFO: Got endpoints: latency-svc-bvt48 [747.981613ms] Jun 3 11:34:55.624: INFO: Created: latency-svc-qmm92 Jun 3 11:34:55.667: INFO: Got endpoints: latency-svc-5m788 [749.281242ms] Jun 3 11:34:55.679: INFO: Created: latency-svc-s5tvg Jun 3 11:34:55.716: INFO: Got endpoints: latency-svc-nq8sn [747.968887ms] Jun 3 11:34:55.721: INFO: Created: latency-svc-4jxqp Jun 3 11:34:55.767: INFO: Got endpoints: latency-svc-btvv5 [749.261904ms] Jun 3 11:34:55.773: INFO: Created: latency-svc-5z8mq Jun 3 11:34:55.817: INFO: Got endpoints: latency-svc-wq4fv [750.157776ms] Jun 3 11:34:55.824: INFO: Created: latency-svc-784md Jun 3 11:34:55.868: INFO: Got endpoints: latency-svc-mlrrh [750.700167ms] Jun 3 11:34:55.875: INFO: Created: latency-svc-jn4mb Jun 3 11:34:55.919: INFO: Got endpoints: latency-svc-8rtzg [751.454222ms] Jun 3 11:34:55.926: INFO: Created: latency-svc-jgnqc Jun 3 11:34:55.968: INFO: Got endpoints: latency-svc-hmp8p [751.726925ms] Jun 3 11:34:55.976: INFO: Created: latency-svc-gx78g Jun 3 11:34:56.018: INFO: Got endpoints: latency-svc-pzqpg [752.200027ms] Jun 3 11:34:56.026: INFO: Created: latency-svc-nh8hp Jun 3 11:34:56.067: INFO: Got endpoints: latency-svc-28rpf [748.738364ms] Jun 3 11:34:56.075: INFO: Created: latency-svc-qcltm Jun 3 11:34:56.117: INFO: Got endpoints: latency-svc-bc6h7 [747.671804ms] Jun 3 11:34:56.124: INFO: Created: latency-svc-fsw8z Jun 3 11:34:56.167: INFO: Got endpoints: latency-svc-frkxr [749.917704ms] Jun 3 11:34:56.175: INFO: Created: latency-svc-ltnt7 Jun 3 11:34:56.216: INFO: Got endpoints: latency-svc-w4t4x [747.952628ms] Jun 3 11:34:56.223: INFO: Created: latency-svc-9dtjj Jun 3 11:34:56.268: INFO: Got endpoints: latency-svc-kl4gh [751.186705ms] Jun 3 11:34:56.276: INFO: Created: latency-svc-xntbn Jun 3 11:34:56.317: INFO: Got endpoints: latency-svc-rpwv8 [750.153118ms] Jun 3 11:34:56.379: INFO: Got endpoints: latency-svc-qmm92 [762.231031ms] Jun 3 11:34:56.383: INFO: Created: latency-svc-s5b5w Jun 3 11:34:56.389: INFO: Created: latency-svc-ppfvz Jun 3 11:34:56.416: INFO: Got endpoints: latency-svc-s5tvg [749.129127ms] Jun 3 11:34:56.423: INFO: Created: latency-svc-hgxld Jun 3 11:34:56.467: INFO: Got endpoints: latency-svc-4jxqp [750.87196ms] Jun 3 11:34:56.474: INFO: Created: latency-svc-cj4lt Jun 3 11:34:56.519: INFO: Got endpoints: latency-svc-5z8mq [752.103837ms] Jun 3 11:34:56.527: INFO: Created: latency-svc-zw6kh Jun 3 11:34:56.568: INFO: Got endpoints: latency-svc-784md [750.641706ms] Jun 3 11:34:56.575: INFO: Created: latency-svc-qff2t Jun 3 11:34:56.618: INFO: Got endpoints: latency-svc-jn4mb [749.948287ms] Jun 3 11:34:56.625: INFO: Created: latency-svc-t5n7p Jun 3 11:34:56.667: INFO: Got endpoints: latency-svc-jgnqc [747.927647ms] Jun 3 11:34:56.674: INFO: Created: latency-svc-qhvb4 Jun 3 11:34:56.779: INFO: Got endpoints: latency-svc-gx78g [810.949922ms] Jun 3 11:34:56.779: INFO: Got endpoints: latency-svc-nh8hp [761.42109ms] Jun 3 11:34:56.786: INFO: Created: latency-svc-477t4 Jun 3 11:34:56.789: INFO: Created: latency-svc-jd8n6 Jun 3 11:34:56.817: INFO: Got endpoints: latency-svc-qcltm [749.601561ms] Jun 3 11:34:56.824: INFO: Created: latency-svc-cn9gh Jun 3 11:34:56.867: INFO: Got endpoints: latency-svc-fsw8z [750.609137ms] Jun 3 11:34:56.875: INFO: Created: latency-svc-7k6hh Jun 3 11:34:56.918: INFO: Got endpoints: latency-svc-ltnt7 [750.710007ms] Jun 3 11:34:56.925: INFO: Created: latency-svc-wh8qp Jun 3 11:34:56.970: INFO: Got endpoints: latency-svc-9dtjj [752.978414ms] Jun 3 11:34:56.977: INFO: Created: latency-svc-fn4rh Jun 3 11:34:57.017: INFO: Got endpoints: latency-svc-xntbn [749.008885ms] Jun 3 11:34:57.025: INFO: Created: latency-svc-42v6p Jun 3 11:34:57.068: INFO: Got endpoints: latency-svc-s5b5w [750.398692ms] Jun 3 11:34:57.075: INFO: Created: latency-svc-w675k Jun 3 11:34:57.117: INFO: Got endpoints: latency-svc-ppfvz [737.572961ms] Jun 3 11:34:57.124: INFO: Created: latency-svc-nj5m9 Jun 3 11:34:57.167: INFO: Got endpoints: latency-svc-hgxld [750.758109ms] Jun 3 11:34:57.174: INFO: Created: latency-svc-j2k4v Jun 3 11:34:57.217: INFO: Got endpoints: latency-svc-cj4lt [750.039548ms] Jun 3 11:34:57.224: INFO: Created: latency-svc-52chs Jun 3 11:34:57.269: INFO: Got endpoints: latency-svc-zw6kh [749.954041ms] Jun 3 11:34:57.275: INFO: Created: latency-svc-qdgzt Jun 3 11:34:57.325: INFO: Got endpoints: latency-svc-qff2t [756.578289ms] Jun 3 11:34:57.358: INFO: Created: latency-svc-qclkr Jun 3 11:34:57.366: INFO: Got endpoints: latency-svc-t5n7p [748.022577ms] Jun 3 11:34:57.372: INFO: Created: latency-svc-c8xxg Jun 3 11:34:57.480: INFO: Got endpoints: latency-svc-qhvb4 [812.892218ms] Jun 3 11:34:57.480: INFO: Got endpoints: latency-svc-477t4 [700.543342ms] Jun 3 11:34:57.492: INFO: Created: latency-svc-gt5kk Jun 3 11:34:57.494: INFO: Created: latency-svc-sfmrr Jun 3 11:34:57.517: INFO: Got endpoints: latency-svc-jd8n6 [737.795081ms] Jun 3 11:34:57.523: INFO: Created: latency-svc-4b556 Jun 3 11:34:57.568: INFO: Got endpoints: latency-svc-cn9gh [750.8499ms] Jun 3 11:34:57.574: INFO: Created: latency-svc-lsjtp Jun 3 11:34:57.684: INFO: Got endpoints: latency-svc-7k6hh [816.838049ms] Jun 3 11:34:57.684: INFO: Got endpoints: latency-svc-wh8qp [766.231771ms] Jun 3 11:34:57.781: INFO: Created: latency-svc-r9hbs Jun 3 11:34:57.781: INFO: Got endpoints: latency-svc-fn4rh [811.30145ms] Jun 3 11:34:57.781: INFO: Got endpoints: latency-svc-42v6p [763.410345ms] Jun 3 11:34:57.788: INFO: Created: latency-svc-85dgb Jun 3 11:34:57.791: INFO: Created: latency-svc-br5lx Jun 3 11:34:57.793: INFO: Created: latency-svc-xvwjj Jun 3 11:34:57.817: INFO: Got endpoints: latency-svc-w675k [749.187444ms] Jun 3 11:34:57.824: INFO: Created: latency-svc-jjjkd Jun 3 11:34:57.867: INFO: Got endpoints: latency-svc-nj5m9 [750.584273ms] Jun 3 11:34:57.874: INFO: Created: latency-svc-6nr95 Jun 3 11:34:57.918: INFO: Got endpoints: latency-svc-j2k4v [750.828686ms] Jun 3 11:34:57.926: INFO: Created: latency-svc-9kmnv Jun 3 11:34:57.967: INFO: Got endpoints: latency-svc-52chs [750.050385ms] Jun 3 11:34:57.974: INFO: Created: latency-svc-fq2rd Jun 3 11:34:58.017: INFO: Got endpoints: latency-svc-qdgzt [748.502596ms] Jun 3 11:34:58.025: INFO: Created: latency-svc-gkb5c Jun 3 11:34:58.068: INFO: Got endpoints: latency-svc-qclkr [742.877558ms] Jun 3 11:34:58.076: INFO: Created: latency-svc-q6w4n Jun 3 11:34:58.118: INFO: Got endpoints: latency-svc-c8xxg [752.046458ms] Jun 3 11:34:58.126: INFO: Created: latency-svc-k7hzf Jun 3 11:34:58.168: INFO: Got endpoints: latency-svc-gt5kk [687.882438ms] Jun 3 11:34:58.175: INFO: Created: latency-svc-k297z Jun 3 11:34:58.218: INFO: Got endpoints: latency-svc-sfmrr [738.105884ms] Jun 3 11:34:58.226: INFO: Created: latency-svc-czdhh Jun 3 11:34:58.268: INFO: Got endpoints: latency-svc-4b556 [751.089475ms] Jun 3 11:34:58.276: INFO: Created: latency-svc-9v6wf Jun 3 11:34:58.381: INFO: Got endpoints: latency-svc-lsjtp [812.875844ms] Jun 3 11:34:58.381: INFO: Got endpoints: latency-svc-r9hbs [695.792362ms] Jun 3 11:34:58.389: INFO: Created: latency-svc-4mxrv Jun 3 11:34:58.393: INFO: Created: latency-svc-j2rc8 Jun 3 11:34:58.418: INFO: Got endpoints: latency-svc-85dgb [733.207366ms] Jun 3 11:34:58.467: INFO: Got endpoints: latency-svc-br5lx [686.300316ms] Jun 3 11:34:58.518: INFO: Got endpoints: latency-svc-xvwjj [736.741577ms] Jun 3 11:34:58.568: INFO: Got endpoints: latency-svc-jjjkd [750.843116ms] Jun 3 11:34:58.618: INFO: Got endpoints: latency-svc-6nr95 [750.634352ms] Jun 3 11:34:58.668: INFO: Got endpoints: latency-svc-9kmnv [750.015782ms] Jun 3 11:34:58.717: INFO: Got endpoints: latency-svc-fq2rd [749.598166ms] Jun 3 11:34:58.770: INFO: Got endpoints: latency-svc-gkb5c [752.064039ms] Jun 3 11:34:58.817: INFO: Got endpoints: latency-svc-q6w4n [748.825291ms] Jun 3 11:34:58.867: INFO: Got endpoints: latency-svc-k7hzf [748.723849ms] Jun 3 11:34:58.917: INFO: Got endpoints: latency-svc-k297z [749.355389ms] Jun 3 11:34:58.968: INFO: Got endpoints: latency-svc-czdhh [749.50853ms] Jun 3 11:34:59.079: INFO: Got endpoints: latency-svc-9v6wf [810.944611ms] Jun 3 11:34:59.079: INFO: Got endpoints: latency-svc-4mxrv [698.461517ms] Jun 3 11:34:59.118: INFO: Got endpoints: latency-svc-j2rc8 [736.369192ms] Jun 3 11:34:59.118: INFO: Latencies: [9.87542ms 13.238305ms 13.788118ms 17.976461ms 20.026816ms 22.519319ms 25.303108ms 28.068618ms 29.609268ms 31.997717ms 32.091352ms 32.289834ms 32.521209ms 32.539536ms 32.617443ms 32.715961ms 32.77836ms 32.977599ms 34.038152ms 34.184118ms 34.228571ms 34.367523ms 34.552025ms 34.813471ms 36.049984ms 36.078839ms 36.464436ms 39.095456ms 40.022738ms 43.131053ms 52.583313ms 101.730529ms 150.059678ms 197.625536ms 248.495338ms 293.371425ms 340.664561ms 389.27356ms 435.933839ms 484.481087ms 533.558889ms 580.393101ms 627.903078ms 676.846872ms 686.300316ms 687.882438ms 689.044239ms 695.792362ms 698.461517ms 700.543342ms 728.567243ms 733.207366ms 736.369192ms 736.741577ms 737.572961ms 737.795081ms 738.105884ms 739.928326ms 742.877558ms 745.17734ms 746.019982ms 746.489147ms 746.802593ms 747.374382ms 747.671804ms 747.810053ms 747.927647ms 747.952628ms 747.968887ms 747.981613ms 748.022577ms 748.32228ms 748.360345ms 748.396963ms 748.502596ms 748.6822ms 748.70382ms 748.723849ms 748.738364ms 748.789414ms 748.797074ms 748.79845ms 748.825291ms 748.977167ms 748.985013ms 749.008885ms 749.028459ms 749.065772ms 749.077887ms 749.129127ms 749.176223ms 749.185752ms 749.187444ms 749.197824ms 749.201704ms 749.217395ms 749.222213ms 749.259839ms 749.261904ms 749.281242ms 749.307528ms 749.318685ms 749.355389ms 749.404883ms 749.438154ms 749.44224ms 749.491221ms 749.50853ms 749.51534ms 749.571192ms 749.575615ms 749.598166ms 749.601561ms 749.778069ms 749.8008ms 749.835173ms 749.914135ms 749.917704ms 749.91777ms 749.941698ms 749.948287ms 749.954041ms 749.957592ms 749.992355ms 750.015782ms 750.039548ms 750.050385ms 750.069655ms 750.153118ms 750.157776ms 750.175987ms 750.190436ms 750.191213ms 750.205281ms 750.23313ms 750.25726ms 750.338938ms 750.343748ms 750.398692ms 750.425648ms 750.525393ms 750.542322ms 750.584273ms 750.589726ms 750.597367ms 750.609137ms 750.634352ms 750.641706ms 750.697822ms 750.698794ms 750.700167ms 750.710007ms 750.750839ms 750.758109ms 750.779467ms 750.798006ms 750.812216ms 750.828686ms 750.835481ms 750.843116ms 750.8499ms 750.87196ms 751.089475ms 751.146225ms 751.171216ms 751.179455ms 751.180699ms 751.186705ms 751.210854ms 751.231235ms 751.248869ms 751.308553ms 751.318902ms 751.345041ms 751.387936ms 751.440851ms 751.454222ms 751.474263ms 751.546969ms 751.555077ms 751.726925ms 752.046458ms 752.064039ms 752.103837ms 752.200027ms 752.971955ms 752.978414ms 756.578289ms 759.961986ms 761.42109ms 762.231031ms 763.410345ms 766.231771ms 809.864941ms 810.944611ms 810.949922ms 811.30145ms 812.875844ms 812.892218ms 816.838049ms] Jun 3 11:34:59.118: INFO: 50 %ile: 749.307528ms Jun 3 11:34:59.118: INFO: 90 %ile: 751.726925ms Jun 3 11:34:59.118: INFO: 99 %ile: 812.892218ms Jun 3 11:34:59.118: INFO: Total sample count: 200 [AfterEach] [sig-network] Service endpoints latency /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:34:59.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svc-latency-2880" for this suite. • [SLOW TEST:11.770 seconds] [sig-network] Service endpoints latency /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should not be very high [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] Service endpoints latency should not be very high [Conformance]","total":-1,"completed":36,"skipped":553,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:55.775: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated [AfterEach] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:02.813: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-7035" for this suite. • [SLOW TEST:7.048 seconds] [sig-api-machinery] ResourceQuota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","total":-1,"completed":25,"skipped":502,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:52.288: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:241 [It] should check if kubectl can dry-run update Pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: running the image k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 Jun 3 11:34:52.316: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7069 run e2e-test-httpd-pod --image=k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 --labels=run=e2e-test-httpd-pod' Jun 3 11:34:52.451: INFO: stderr: "" Jun 3 11:34:52.451: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: replace the image in the pod with server-side dry-run Jun 3 11:34:52.452: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7069 patch pod e2e-test-httpd-pod -p {"spec":{"containers":[{"name": "e2e-test-httpd-pod","image": "k8s.gcr.io/e2e-test-images/busybox:1.29-1"}]}} --dry-run=server' Jun 3 11:34:52.854: INFO: stderr: "" Jun 3 11:34:52.854: INFO: stdout: "pod/e2e-test-httpd-pod patched\n" STEP: verifying the pod e2e-test-httpd-pod has the right image k8s.gcr.io/e2e-test-images/httpd:2.4.38-1 Jun 3 11:34:52.857: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=kubectl-7069 delete pods e2e-test-httpd-pod' Jun 3 11:35:05.049: INFO: stderr: "" Jun 3 11:35:05.049: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:05.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7069" for this suite. • [SLOW TEST:12.769 seconds] [sig-cli] Kubectl client /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl server-side dry-run /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:903 should check if kubectl can dry-run update Pods [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance]","total":-1,"completed":46,"skipped":720,"failed":0} S ------------------------------ [BeforeEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:02.832: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating configMap with name configmap-test-volume-e99d372b-f85e-4394-af3a-2209f53b8047 STEP: Creating a pod to test consume configMaps Jun 3 11:35:02.872: INFO: Waiting up to 5m0s for pod "pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521" in namespace "configmap-7325" to be "Succeeded or Failed" Jun 3 11:35:02.874: INFO: Pod "pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521": Phase="Pending", Reason="", readiness=false. Elapsed: 2.363375ms Jun 3 11:35:04.878: INFO: Pod "pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521": Phase="Running", Reason="", readiness=true. Elapsed: 2.005830941s Jun 3 11:35:06.883: INFO: Pod "pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010712921s STEP: Saw pod success Jun 3 11:35:06.883: INFO: Pod "pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521" satisfied condition "Succeeded or Failed" Jun 3 11:35:06.887: INFO: Trying to get logs from node v1.21-worker pod pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521 container agnhost-container: STEP: delete the pod Jun 3 11:35:06.901: INFO: Waiting for pod pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521 to disappear Jun 3 11:35:06.906: INFO: Pod pod-configmaps-1c07d505-209d-47d3-ac3c-2bdb3d604521 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:06.906: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7325" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":26,"skipped":506,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:58.910: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: Gathering metrics W0603 11:34:04.974424 30 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Jun 3 11:35:07.017: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:07.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-7906" for this suite. • [SLOW TEST:68.126 seconds] [sig-api-machinery] Garbage collector /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","total":-1,"completed":10,"skipped":219,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:07.051: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41 [It] should provide container's cpu limit [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test downward API volume plugin Jun 3 11:35:07.083: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5e0fc3ab-f5e0-4ba8-86f5-ea6e15a3792f" in namespace "downward-api-9249" to be "Succeeded or Failed" Jun 3 11:35:07.085: INFO: Pod "downwardapi-volume-5e0fc3ab-f5e0-4ba8-86f5-ea6e15a3792f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.363534ms Jun 3 11:35:09.092: INFO: Pod "downwardapi-volume-5e0fc3ab-f5e0-4ba8-86f5-ea6e15a3792f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.008407552s STEP: Saw pod success Jun 3 11:35:09.092: INFO: Pod "downwardapi-volume-5e0fc3ab-f5e0-4ba8-86f5-ea6e15a3792f" satisfied condition "Succeeded or Failed" Jun 3 11:35:09.094: INFO: Trying to get logs from node v1.21-worker pod downwardapi-volume-5e0fc3ab-f5e0-4ba8-86f5-ea6e15a3792f container client-container: STEP: delete the pod Jun 3 11:35:09.106: INFO: Waiting for pod downwardapi-volume-5e0fc3ab-f5e0-4ba8-86f5-ea6e15a3792f to disappear Jun 3 11:35:09.108: INFO: Pod downwardapi-volume-5e0fc3ab-f5e0-4ba8-86f5-ea6e15a3792f no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:09.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-9249" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":-1,"completed":11,"skipped":228,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:59.145: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD preserving unknown fields in an embedded object [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:34:59.176: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Jun 3 11:35:03.802: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8453 --namespace=crd-publish-openapi-8453 create -f -' Jun 3 11:35:04.384: INFO: stderr: "" Jun 3 11:35:04.384: INFO: stdout: "e2e-test-crd-publish-openapi-9137-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" Jun 3 11:35:04.384: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8453 --namespace=crd-publish-openapi-8453 delete e2e-test-crd-publish-openapi-9137-crds test-cr' Jun 3 11:35:04.511: INFO: stderr: "" Jun 3 11:35:04.511: INFO: stdout: "e2e-test-crd-publish-openapi-9137-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" Jun 3 11:35:04.511: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8453 --namespace=crd-publish-openapi-8453 apply -f -' Jun 3 11:35:04.998: INFO: stderr: "" Jun 3 11:35:04.998: INFO: stdout: "e2e-test-crd-publish-openapi-9137-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" Jun 3 11:35:04.998: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8453 --namespace=crd-publish-openapi-8453 delete e2e-test-crd-publish-openapi-9137-crds test-cr' Jun 3 11:35:05.127: INFO: stderr: "" Jun 3 11:35:05.127: INFO: stdout: "e2e-test-crd-publish-openapi-9137-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR Jun 3 11:35:05.127: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8453 explain e2e-test-crd-publish-openapi-9137-crds' Jun 3 11:35:05.515: INFO: stderr: "" Jun 3 11:35:05.515: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-9137-crd\nVERSION: crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t<>\n Specification of Waldo\n\n status\t\n Status of Waldo\n\n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:10.212: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-8453" for this suite. • [SLOW TEST:11.075 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD preserving unknown fields in an embedded object [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":-1,"completed":37,"skipped":558,"failed":0} SSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:54.549: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:34:55.269: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Jun 3 11:34:57.279: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316895, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316895, loc:(*time.Location)(0x9dc0820)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316895, loc:(*time.Location)(0x9dc0820)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63758316895, loc:(*time.Location)(0x9dc0820)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:35:00.379: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] patching/updating a validating webhook should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a validating webhook configuration Jun 3 11:35:10.399: INFO: Waiting for webhook configuration to be ready... STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Updating a validating webhook configuration's rules to not include the create operation STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Patching a validating webhook configuration's rules to include the create operation STEP: Creating a configMap that does not comply to the validation webhook rules [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:10.556: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-6050" for this suite. STEP: Destroying namespace "webhook-6050-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:16.046 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 patching/updating a validating webhook should work [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","total":-1,"completed":51,"skipped":667,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-api-machinery] server version /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:10.605: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename server-version STEP: Waiting for a default service account to be provisioned in namespace [It] should find the server version [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Request ServerVersion STEP: Confirm major version Jun 3 11:35:10.636: INFO: Major version: 1 STEP: Confirm minor version Jun 3 11:35:10.636: INFO: cleanMinorVersion: 21 Jun 3 11:35:10.636: INFO: Minor version: 21 [AfterEach] [sig-api-machinery] server version /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:10.636: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "server-version-6756" for this suite. • ------------------------------ {"msg":"PASSED [sig-api-machinery] server version should find the server version [Conformance]","total":-1,"completed":52,"skipped":672,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:10.656: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name secret-test-map-be566bc0-1d43-4746-9cf4-b8f44858bbe9 STEP: Creating a pod to test consume secrets Jun 3 11:35:10.693: INFO: Waiting up to 5m0s for pod "pod-secrets-03319da3-3b87-430b-9600-b2e91cc27c72" in namespace "secrets-7417" to be "Succeeded or Failed" Jun 3 11:35:10.695: INFO: Pod "pod-secrets-03319da3-3b87-430b-9600-b2e91cc27c72": Phase="Pending", Reason="", readiness=false. Elapsed: 2.607636ms Jun 3 11:35:12.699: INFO: Pod "pod-secrets-03319da3-3b87-430b-9600-b2e91cc27c72": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.006362208s STEP: Saw pod success Jun 3 11:35:12.699: INFO: Pod "pod-secrets-03319da3-3b87-430b-9600-b2e91cc27c72" satisfied condition "Succeeded or Failed" Jun 3 11:35:12.703: INFO: Trying to get logs from node v1.21-worker2 pod pod-secrets-03319da3-3b87-430b-9600-b2e91cc27c72 container secret-volume-test: STEP: delete the pod Jun 3 11:35:12.718: INFO: Waiting for pod pod-secrets-03319da3-3b87-430b-9600-b2e91cc27c72 to disappear Jun 3 11:35:12.721: INFO: Pod pod-secrets-03319da3-3b87-430b-9600-b2e91cc27c72 no longer exists [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:12.721: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-7417" for this suite. • ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":-1,"completed":53,"skipped":678,"failed":0} SSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:42.810: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 Jun 3 11:34:42.849: INFO: created pod Jun 3 11:34:42.849: INFO: Waiting up to 5m0s for pod "oidc-discovery-validator" in namespace "svcaccounts-1332" to be "Succeeded or Failed" Jun 3 11:34:42.852: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 2.720409ms Jun 3 11:34:44.855: INFO: Pod "oidc-discovery-validator": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.006493655s STEP: Saw pod success Jun 3 11:34:44.855: INFO: Pod "oidc-discovery-validator" satisfied condition "Succeeded or Failed" Jun 3 11:35:14.856: INFO: polling logs Jun 3 11:35:14.864: INFO: Pod logs: 2021/06/03 11:34:43 OK: Got token 2021/06/03 11:34:43 validating with in-cluster discovery 2021/06/03 11:34:43 OK: got issuer https://kubernetes.default.svc.cluster.local 2021/06/03 11:34:43 Full, not-validated claims: openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-1332:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1622720683, NotBefore:1622720083, IssuedAt:1622720083, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-1332", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"5c22e105-47a0-4a77-a8ab-b5aa65eb11ed"}}} 2021/06/03 11:34:43 OK: Constructed OIDC provider for issuer https://kubernetes.default.svc.cluster.local 2021/06/03 11:34:43 OK: Validated signature on JWT 2021/06/03 11:34:43 OK: Got valid claims from token! 2021/06/03 11:34:43 Full, validated claims: &openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-1332:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1622720683, NotBefore:1622720083, IssuedAt:1622720083, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-1332", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"5c22e105-47a0-4a77-a8ab-b5aa65eb11ed"}}} Jun 3 11:35:14.864: INFO: completed pod [AfterEach] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:14.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-1332" for this suite. • [SLOW TEST:32.073 seconds] [sig-auth] ServiceAccounts /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance]","total":-1,"completed":19,"skipped":286,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:10.236: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Jun 3 11:35:10.833: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Jun 3 11:35:13.853: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 Jun 3 11:35:14.852: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 Jun 3 11:35:15.853: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 Jun 3 11:35:16.852: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 Jun 3 11:35:17.853: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny attaching pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Registering the webhook via the AdmissionRegistration API STEP: create a pod STEP: 'kubectl attach' the pod, should be denied by the webhook Jun 3 11:35:19.886: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=webhook-1775 attach --namespace=webhook-1775 to-be-attached-pod -i -c=container1' Jun 3 11:35:20.041: INFO: rc: 1 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:20.045: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-1775" for this suite. STEP: Destroying namespace "webhook-1775-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:9.841 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny attaching pod [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","total":-1,"completed":38,"skipped":567,"failed":0} SSSSSS ------------------------------ [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:14.952: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41 [It] should update annotations on modification [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating the pod Jun 3 11:35:15.277: INFO: The status of Pod annotationupdate7d7d5580-0ff9-4bff-8596-e4303d222994 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:35:17.282: INFO: The status of Pod annotationupdate7d7d5580-0ff9-4bff-8596-e4303d222994 is Running (Ready = true) Jun 3 11:35:17.803: INFO: Successfully updated pod "annotationupdate7d7d5580-0ff9-4bff-8596-e4303d222994" [AfterEach] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:21.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7887" for this suite. • [SLOW TEST:6.881 seconds] [sig-storage] Projected downwardAPI /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 should update annotations on modification [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","total":-1,"completed":20,"skipped":324,"failed":0} SSSS ------------------------------ [BeforeEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:20.088: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a pod to test override arguments Jun 3 11:35:20.123: INFO: Waiting up to 5m0s for pod "client-containers-c8797196-6952-418d-9ab1-e5e1083120ab" in namespace "containers-5572" to be "Succeeded or Failed" Jun 3 11:35:20.125: INFO: Pod "client-containers-c8797196-6952-418d-9ab1-e5e1083120ab": Phase="Pending", Reason="", readiness=false. Elapsed: 2.595585ms Jun 3 11:35:22.130: INFO: Pod "client-containers-c8797196-6952-418d-9ab1-e5e1083120ab": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.006994704s STEP: Saw pod success Jun 3 11:35:22.130: INFO: Pod "client-containers-c8797196-6952-418d-9ab1-e5e1083120ab" satisfied condition "Succeeded or Failed" Jun 3 11:35:22.133: INFO: Trying to get logs from node v1.21-worker2 pod client-containers-c8797196-6952-418d-9ab1-e5e1083120ab container agnhost-container: STEP: delete the pod Jun 3 11:35:22.147: INFO: Waiting for pod client-containers-c8797196-6952-418d-9ab1-e5e1083120ab to disappear Jun 3 11:35:22.150: INFO: Pod client-containers-c8797196-6952-418d-9ab1-e5e1083120ab no longer exists [AfterEach] [sig-node] Docker Containers /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:22.150: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-5572" for this suite. • ------------------------------ {"msg":"PASSED [sig-node] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]","total":-1,"completed":39,"skipped":573,"failed":0} SSSSSSS ------------------------------ [BeforeEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:06.985: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:52 STEP: create the container to handle the HTTPGet hook request. Jun 3 11:35:07.056: INFO: The status of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:35:09.061: INFO: The status of Pod pod-handle-http-request is Running (Ready = true) [It] should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: create the pod with lifecycle hook Jun 3 11:35:09.081: INFO: The status of Pod pod-with-prestop-exec-hook is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:35:11.085: INFO: The status of Pod pod-with-prestop-exec-hook is Running (Ready = true) STEP: delete the pod with lifecycle hook Jun 3 11:35:11.091: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:11.094: INFO: Pod pod-with-prestop-exec-hook still exists Jun 3 11:35:13.095: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:13.099: INFO: Pod pod-with-prestop-exec-hook still exists Jun 3 11:35:15.094: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:15.178: INFO: Pod pod-with-prestop-exec-hook still exists Jun 3 11:35:17.095: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:17.099: INFO: Pod pod-with-prestop-exec-hook still exists Jun 3 11:35:19.096: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:19.100: INFO: Pod pod-with-prestop-exec-hook still exists Jun 3 11:35:21.095: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:21.099: INFO: Pod pod-with-prestop-exec-hook still exists Jun 3 11:35:23.095: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:23.100: INFO: Pod pod-with-prestop-exec-hook still exists Jun 3 11:35:25.094: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jun 3 11:35:25.097: INFO: Pod pod-with-prestop-exec-hook no longer exists STEP: check prestop hook [AfterEach] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:25.103: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-7970" for this suite. • [SLOW TEST:18.127 seconds] [sig-node] Container Lifecycle Hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 when create a pod with lifecycle hook /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:43 should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":-1,"completed":27,"skipped":549,"failed":0} Jun 3 11:35:25.113: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:12.755: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of different groups [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation Jun 3 11:35:12.786: INFO: >>> kubeConfig: /root/.kube/config Jun 3 11:35:17.410: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:35.556: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-3674" for this suite. • [SLOW TEST:22.809 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of different groups [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","total":-1,"completed":54,"skipped":690,"failed":0} Jun 3 11:35:35.566: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:05.061: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for pods for Subdomain [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6906.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-6906.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6906.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:35:07.114: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.117: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.120: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.123: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.134: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.137: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.140: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.143: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:07.149: INFO: Lookups using dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local] Jun 3 11:35:12.153: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.156: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.159: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.163: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.172: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.175: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.178: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.181: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:12.188: INFO: Lookups using dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local] Jun 3 11:35:17.154: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.158: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.161: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.164: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.174: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.178: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.182: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.185: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:17.192: INFO: Lookups using dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local] Jun 3 11:35:22.154: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.157: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.160: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.164: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.173: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.176: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.179: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.183: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:22.189: INFO: Lookups using dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local] Jun 3 11:35:27.154: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.158: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.161: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.165: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.176: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.179: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.182: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.185: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:27.191: INFO: Lookups using dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local] Jun 3 11:35:32.154: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.156: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.159: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.162: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.171: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.174: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.177: INFO: Unable to read jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.180: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local from pod dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be: the server could not find the requested resource (get pods dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be) Jun 3 11:35:32.186: INFO: Lookups using dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local wheezy_udp@dns-test-service-2.dns-6906.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6906.svc.cluster.local jessie_udp@dns-test-service-2.dns-6906.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6906.svc.cluster.local] Jun 3 11:35:37.197: INFO: DNS probes using dns-6906/dns-test-f8c63726-7176-4ad1-8995-3cd1c38601be succeeded STEP: deleting the pod STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:37.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-6906" for this suite. • [SLOW TEST:32.170 seconds] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should provide DNS for pods for Subdomain [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":-1,"completed":47,"skipped":721,"failed":0} Jun 3 11:35:37.233: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:09.144: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6528 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6528;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6528 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6528;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6528.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6528.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6528.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6528.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-6528.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-6528.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6528.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 44.21.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.21.44_udp@PTR;check="$$(dig +tcp +noall +answer +search 44.21.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.21.44_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6528 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6528;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6528 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6528;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6528.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6528.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6528.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6528.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-6528.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-6528.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6528.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-6528.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6528.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 44.21.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.21.44_udp@PTR;check="$$(dig +tcp +noall +answer +search 44.21.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.21.44_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Jun 3 11:35:11.205: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.209: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.213: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.217: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.220: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.223: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.227: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.232: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.256: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.259: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.262: INFO: Unable to read jessie_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.266: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.269: INFO: Unable to read jessie_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.273: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.277: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.281: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:11.303: INFO: Lookups using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6528 wheezy_tcp@dns-test-service.dns-6528 wheezy_udp@dns-test-service.dns-6528.svc wheezy_tcp@dns-test-service.dns-6528.svc wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6528 jessie_tcp@dns-test-service.dns-6528 jessie_udp@dns-test-service.dns-6528.svc jessie_tcp@dns-test-service.dns-6528.svc jessie_udp@_http._tcp.dns-test-service.dns-6528.svc jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc] Jun 3 11:35:16.308: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.312: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.316: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.319: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.323: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.327: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.331: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.334: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.359: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.362: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.366: INFO: Unable to read jessie_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.369: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.372: INFO: Unable to read jessie_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.376: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.379: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.383: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:16.406: INFO: Lookups using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6528 wheezy_tcp@dns-test-service.dns-6528 wheezy_udp@dns-test-service.dns-6528.svc wheezy_tcp@dns-test-service.dns-6528.svc wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6528 jessie_tcp@dns-test-service.dns-6528 jessie_udp@dns-test-service.dns-6528.svc jessie_tcp@dns-test-service.dns-6528.svc jessie_udp@_http._tcp.dns-test-service.dns-6528.svc jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc] Jun 3 11:35:21.309: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.314: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.317: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.321: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.324: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.328: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.331: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.335: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.360: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.364: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.372: INFO: Unable to read jessie_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.375: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.379: INFO: Unable to read jessie_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.383: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.387: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.390: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:21.412: INFO: Lookups using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6528 wheezy_tcp@dns-test-service.dns-6528 wheezy_udp@dns-test-service.dns-6528.svc wheezy_tcp@dns-test-service.dns-6528.svc wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6528 jessie_tcp@dns-test-service.dns-6528 jessie_udp@dns-test-service.dns-6528.svc jessie_tcp@dns-test-service.dns-6528.svc jessie_udp@_http._tcp.dns-test-service.dns-6528.svc jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc] Jun 3 11:35:26.308: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.312: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.315: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.318: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.322: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.325: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.328: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.331: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.355: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.359: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.363: INFO: Unable to read jessie_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.366: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.370: INFO: Unable to read jessie_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.373: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.377: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.381: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:26.402: INFO: Lookups using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6528 wheezy_tcp@dns-test-service.dns-6528 wheezy_udp@dns-test-service.dns-6528.svc wheezy_tcp@dns-test-service.dns-6528.svc wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6528 jessie_tcp@dns-test-service.dns-6528 jessie_udp@dns-test-service.dns-6528.svc jessie_tcp@dns-test-service.dns-6528.svc jessie_udp@_http._tcp.dns-test-service.dns-6528.svc jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc] Jun 3 11:35:31.310: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.314: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.318: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.322: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.326: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.330: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.333: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.337: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.364: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.368: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.372: INFO: Unable to read jessie_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.376: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.379: INFO: Unable to read jessie_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.382: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.386: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.390: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:31.412: INFO: Lookups using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6528 wheezy_tcp@dns-test-service.dns-6528 wheezy_udp@dns-test-service.dns-6528.svc wheezy_tcp@dns-test-service.dns-6528.svc wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6528 jessie_tcp@dns-test-service.dns-6528 jessie_udp@dns-test-service.dns-6528.svc jessie_tcp@dns-test-service.dns-6528.svc jessie_udp@_http._tcp.dns-test-service.dns-6528.svc jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc] Jun 3 11:35:36.308: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.312: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.316: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.320: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.324: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.328: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.332: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.336: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.363: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.367: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.370: INFO: Unable to read jessie_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.373: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.377: INFO: Unable to read jessie_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.381: INFO: Unable to read jessie_tcp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.384: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.388: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:36.411: INFO: Lookups using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6528 wheezy_tcp@dns-test-service.dns-6528 wheezy_udp@dns-test-service.dns-6528.svc wheezy_tcp@dns-test-service.dns-6528.svc wheezy_udp@_http._tcp.dns-test-service.dns-6528.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6528.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6528 jessie_tcp@dns-test-service.dns-6528 jessie_udp@dns-test-service.dns-6528.svc jessie_tcp@dns-test-service.dns-6528.svc jessie_udp@_http._tcp.dns-test-service.dns-6528.svc jessie_tcp@_http._tcp.dns-test-service.dns-6528.svc] Jun 3 11:35:41.310: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:41.315: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:41.319: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:41.323: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6528 from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:41.326: INFO: Unable to read wheezy_udp@dns-test-service.dns-6528.svc from pod dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5: the server could not find the requested resource (get pods dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5) Jun 3 11:35:41.418: INFO: Lookups using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6528 wheezy_tcp@dns-test-service.dns-6528 wheezy_udp@dns-test-service.dns-6528.svc] Jun 3 11:35:46.413: INFO: DNS probes using dns-6528/dns-test-383e88cb-355e-4abf-97ed-cbf66e7845f5 succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:35:46.463: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-6528" for this suite. • [SLOW TEST:37.328 seconds] [sig-network] DNS /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23 should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","total":-1,"completed":12,"skipped":245,"failed":0} Jun 3 11:35:46.474: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:35:21.843: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating secret with name s-test-opt-del-1035c9e1-7354-4c35-81b3-4edf5e62b881 STEP: Creating secret with name s-test-opt-upd-6addbf6b-ba27-4bd6-84ff-5b1783a4302c STEP: Creating the pod Jun 3 11:35:21.897: INFO: The status of Pod pod-secrets-c8b9852e-697b-450a-ae68-cfbd645346a6 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:35:23.902: INFO: The status of Pod pod-secrets-c8b9852e-697b-450a-ae68-cfbd645346a6 is Pending, waiting for it to be Running (with Ready = true) Jun 3 11:35:25.902: INFO: The status of Pod pod-secrets-c8b9852e-697b-450a-ae68-cfbd645346a6 is Running (Ready = true) STEP: Deleting secret s-test-opt-del-1035c9e1-7354-4c35-81b3-4edf5e62b881 STEP: Updating secret s-test-opt-upd-6addbf6b-ba27-4bd6-84ff-5b1783a4302c STEP: Creating secret with name s-test-opt-create-8219a34a-2a13-469f-b799-7c79e052b182 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:36:54.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-6597" for this suite. • [SLOW TEST:92.694 seconds] [sig-storage] Secrets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":21,"skipped":328,"failed":0} Jun 3 11:36:54.539: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:33:25.599: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod liveness-dac573fa-8fa5-4b7f-b166-2c641a1f5705 in namespace container-probe-3324 Jun 3 11:33:33.644: INFO: Started pod liveness-dac573fa-8fa5-4b7f-b166-2c641a1f5705 in namespace container-probe-3324 STEP: checking the pod's current state and verifying that restartCount is present Jun 3 11:33:33.647: INFO: Initial restart count of pod liveness-dac573fa-8fa5-4b7f-b166-2c641a1f5705 is 0 STEP: deleting the pod [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:37:35.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-3324" for this suite. • [SLOW TEST:249.728 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]","total":-1,"completed":23,"skipped":386,"failed":0} Jun 3 11:37:35.330: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:32:19.011: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105 STEP: Creating service test in namespace statefulset-5626 [It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating stateful set ss in namespace statefulset-5626 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-5626 Jun 3 11:32:19.048: INFO: Found 0 stateful pods, waiting for 1 Jun 3 11:32:29.079: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod Jun 3 11:32:29.182: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:32:29.699: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:32:29.699: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:32:29.699: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:32:29.703: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Jun 3 11:32:39.706: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:32:39.706: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:32:39.718: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:32:39.718: INFO: ss-0 v1.21-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:30 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:30 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC }] Jun 3 11:32:39.718: INFO: Jun 3 11:32:39.718: INFO: StatefulSet ss has not reached scale 3, at 1 Jun 3 11:32:40.980: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.996990614s Jun 3 11:32:42.080: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.734429807s Jun 3 11:32:43.085: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.635469502s Jun 3 11:32:44.098: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.629556947s Jun 3 11:32:45.102: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.616182407s Jun 3 11:32:46.107: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.612644738s Jun 3 11:32:47.111: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.607582219s Jun 3 11:32:48.115: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.603804407s Jun 3 11:32:49.120: INFO: Verifying statefulset ss doesn't scale past 3 for another 599.212263ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-5626 Jun 3 11:32:50.127: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:32:50.388: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 3 11:32:50.388: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:32:50.388: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:32:50.388: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:32:50.648: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" Jun 3 11:32:50.648: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:32:50.648: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:32:50.648: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:32:50.898: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" Jun 3 11:32:50.898: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 3 11:32:50.898: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 3 11:32:50.902: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=false Jun 3 11:33:00.908: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:33:00.908: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jun 3 11:33:00.908: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Scale down will not halt with unhealthy stateful pod Jun 3 11:33:00.912: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:33:01.165: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:33:01.166: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:33:01.166: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:33:01.166: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:33:01.397: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:33:01.397: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:33:01.397: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:33:01.397: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 3 11:33:01.644: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 3 11:33:01.644: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 3 11:33:01.644: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 3 11:33:01.644: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:33:01.648: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3 Jun 3 11:33:11.657: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:33:11.657: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:33:11.657: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Jun 3 11:33:11.670: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:11.670: INFO: ss-0 v1.21-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC }] Jun 3 11:33:11.670: INFO: ss-1 v1.21-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:11.670: INFO: ss-2 v1.21-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:11.670: INFO: Jun 3 11:33:11.670: INFO: StatefulSet ss has not reached scale 0, at 3 Jun 3 11:33:12.678: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:12.678: INFO: ss-0 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC }] Jun 3 11:33:12.678: INFO: ss-1 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:12.679: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:12.679: INFO: Jun 3 11:33:12.679: INFO: StatefulSet ss has not reached scale 0, at 3 Jun 3 11:33:13.683: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:13.683: INFO: ss-0 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC }] Jun 3 11:33:13.684: INFO: ss-1 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:13.684: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:13.684: INFO: Jun 3 11:33:13.684: INFO: StatefulSet ss has not reached scale 0, at 3 Jun 3 11:33:14.689: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:14.689: INFO: ss-0 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC }] Jun 3 11:33:14.689: INFO: ss-1 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:14.689: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:14.689: INFO: Jun 3 11:33:14.689: INFO: StatefulSet ss has not reached scale 0, at 3 Jun 3 11:33:15.693: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:15.693: INFO: ss-0 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC }] Jun 3 11:33:15.693: INFO: ss-1 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:15.694: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:15.694: INFO: Jun 3 11:33:15.694: INFO: StatefulSet ss has not reached scale 0, at 3 Jun 3 11:33:16.698: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:16.698: INFO: ss-0 v1.21-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:19 +0000 UTC }] Jun 3 11:33:16.698: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:16.698: INFO: Jun 3 11:33:16.698: INFO: StatefulSet ss has not reached scale 0, at 2 Jun 3 11:33:17.703: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:17.703: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:17.703: INFO: Jun 3 11:33:17.703: INFO: StatefulSet ss has not reached scale 0, at 1 Jun 3 11:33:18.707: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:18.707: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:18.707: INFO: Jun 3 11:33:18.707: INFO: StatefulSet ss has not reached scale 0, at 1 Jun 3 11:33:19.716: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:19.716: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:19.716: INFO: Jun 3 11:33:19.716: INFO: StatefulSet ss has not reached scale 0, at 1 Jun 3 11:33:20.721: INFO: POD NODE PHASE GRACE CONDITIONS Jun 3 11:33:20.721: INFO: ss-2 v1.21-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:33:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-06-03 11:32:39 +0000 UTC }] Jun 3 11:33:20.721: INFO: Jun 3 11:33:20.721: INFO: StatefulSet ss has not reached scale 0, at 1 STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-5626 Jun 3 11:33:21.727: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:33:21.885: INFO: rc: 1 Jun 3 11:33:21.885: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: error: unable to upgrade connection: container not found ("webserver") error: exit status 1 Jun 3 11:33:31.886: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:33:32.019: INFO: rc: 1 Jun 3 11:33:32.019: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:33:42.020: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:33:42.190: INFO: rc: 1 Jun 3 11:33:42.190: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:33:52.190: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:33:52.317: INFO: rc: 1 Jun 3 11:33:52.317: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:34:02.318: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:34:02.441: INFO: rc: 1 Jun 3 11:34:02.441: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:34:12.442: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:34:12.556: INFO: rc: 1 Jun 3 11:34:12.556: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:34:22.557: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:34:22.679: INFO: rc: 1 Jun 3 11:34:22.679: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:34:32.680: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:34:32.803: INFO: rc: 1 Jun 3 11:34:32.803: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:34:42.804: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:34:42.921: INFO: rc: 1 Jun 3 11:34:42.921: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:34:52.922: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:34:53.051: INFO: rc: 1 Jun 3 11:34:53.051: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:35:03.053: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:35:03.177: INFO: rc: 1 Jun 3 11:35:03.177: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:35:13.179: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:35:13.296: INFO: rc: 1 Jun 3 11:35:13.296: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:35:23.298: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:35:23.414: INFO: rc: 1 Jun 3 11:35:23.414: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:35:33.416: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:35:33.535: INFO: rc: 1 Jun 3 11:35:33.535: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:35:43.537: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:35:43.662: INFO: rc: 1 Jun 3 11:35:43.663: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:35:53.664: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:35:53.787: INFO: rc: 1 Jun 3 11:35:53.787: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:36:03.789: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:36:03.914: INFO: rc: 1 Jun 3 11:36:03.914: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:36:13.917: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:36:14.089: INFO: rc: 1 Jun 3 11:36:14.089: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:36:24.089: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:36:24.215: INFO: rc: 1 Jun 3 11:36:24.215: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:36:34.218: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:36:34.344: INFO: rc: 1 Jun 3 11:36:34.344: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:36:44.346: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:36:44.470: INFO: rc: 1 Jun 3 11:36:44.471: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:36:54.471: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:36:54.590: INFO: rc: 1 Jun 3 11:36:54.591: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:37:04.591: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:37:04.710: INFO: rc: 1 Jun 3 11:37:04.710: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:37:14.710: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:37:14.831: INFO: rc: 1 Jun 3 11:37:14.831: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:37:24.832: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:37:24.958: INFO: rc: 1 Jun 3 11:37:24.958: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:37:34.959: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:37:35.080: INFO: rc: 1 Jun 3 11:37:35.080: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:37:45.081: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:37:45.203: INFO: rc: 1 Jun 3 11:37:45.203: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:37:55.204: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:37:55.332: INFO: rc: 1 Jun 3 11:37:55.332: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:38:05.332: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:38:05.458: INFO: rc: 1 Jun 3 11:38:05.458: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:38:15.458: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:38:15.605: INFO: rc: 1 Jun 3 11:38:15.605: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Jun 3 11:38:25.606: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.90:33295 --kubeconfig=/root/.kube/config --namespace=statefulset-5626 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 3 11:38:25.727: INFO: rc: 1 Jun 3 11:38:25.727: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: Jun 3 11:38:25.727: INFO: Scaling statefulset ss to 0 Jun 3 11:38:25.751: INFO: Waiting for statefulset status.replicas updated to 0 [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116 Jun 3 11:38:25.754: INFO: Deleting all statefulset in ns statefulset-5626 Jun 3 11:38:25.757: INFO: Scaling statefulset ss to 0 Jun 3 11:38:25.768: INFO: Waiting for statefulset status.replicas updated to 0 Jun 3 11:38:25.771: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:38:25.788: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-5626" for this suite. • [SLOW TEST:366.786 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95 Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","total":-1,"completed":21,"skipped":363,"failed":0} Jun 3 11:38:25.802: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:45.453: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:54 [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating pod busybox-b37db004-edc4-4fff-a825-5bd9c2824504 in namespace container-probe-134 Jun 3 11:34:51.493: INFO: Started pod busybox-b37db004-edc4-4fff-a825-5bd9c2824504 in namespace container-probe-134 STEP: checking the pod's current state and verifying that restartCount is present Jun 3 11:34:51.496: INFO: Initial restart count of pod busybox-b37db004-edc4-4fff-a825-5bd9c2824504 is 0 STEP: deleting the pod [AfterEach] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:38:52.618: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-134" for this suite. • [SLOW TEST:247.180 seconds] [sig-node] Probing container /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23 should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-node] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":-1,"completed":44,"skipped":803,"failed":0} Jun 3 11:38:52.635: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185 STEP: Creating a kubernetes client Jun 3 11:34:25.316: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:63 W0603 11:34:25.349001 32 warnings.go:70] batch/v1beta1 CronJob is deprecated in v1.21+, unavailable in v1.25+; use batch/v1 CronJob [It] should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 STEP: Creating a ForbidConcurrent cronjob STEP: Ensuring a job is scheduled STEP: Ensuring exactly one is scheduled STEP: Ensuring exactly one running job exists by listing jobs explicitly STEP: Ensuring no more jobs are scheduled STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186 Jun 3 11:40:01.382: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-8493" for this suite. • [SLOW TEST:336.079 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow] [Conformance]","total":-1,"completed":32,"skipped":786,"failed":0} Jun 3 11:40:01.397: INFO: Running AfterSuite actions on all nodes Jun 3 11:35:22.174: INFO: Running AfterSuite actions on all nodes Jun 3 11:40:01.427: INFO: Running AfterSuite actions on node 1 Jun 3 11:40:01.427: INFO: Skipping dumping logs from cluster Ran 320 of 5771 Specs in 850.221 seconds SUCCESS! -- 320 Passed | 0 Failed | 0 Pending | 5451 Skipped Ginkgo ran 1 suite in 14m12.044614232s Test Suite Passed