<?xml version="1.0" encoding="UTF-8"?>
  <testsuite name="Kubernetes e2e suite" tests="2" failures="2" errors="0" time="1236.375">
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] [Serial] Volume metrics PVController should create unbound pv count metrics for pvc controller after creating pv only" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cluster-lifecycle] [Feature:BootstrapTokens] should sign the new added bootstrap tokens" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:NEG] rolling update backend pods should not cause service disruption" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] client-go should negotiate watch and report errors with accept &#34;application/json&#34;" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] DNS horizontal autoscaling [DisabledForLargeClusters] kube-dns-autoscaler should scale kube-dns pods in both nonfaulty and faulty scenarios" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Security Context should support container.SecurityContext.RunAsUser [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be able to change the type and ports of a service [Slow] [DisabledForLargeClusters]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Security Context should support seccomp default which is unconfined [Feature:Seccomp] [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Socket [Slow] Should fail on mounting socket &#39;asocket&#39; when HostPathType is HostPathFile" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Namespaces [Serial] should delete fast enough (90 percent of 100 namespaces in 150 seconds)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should have cluster metrics [Feature:StackdriverMonitoring]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning DynamicProvisioner [Slow] should provision storage with non-default reclaim policy Retain" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce updated policy [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-bindmounted] Set fsGroup for local volume should set fsGroup for one pod [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] PodSecurityPolicy [Feature:PodSecurityPolicy] should forbid pod creation when no PSP is available" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Ingress API should support creating Ingress API operations [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] Hybrid cluster network for all supported CNIs should have stable networking for Linux and Windows pods" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes [Feature:vsphere][Feature:LabelSelector] [sig-storage] Selector-Label Volume Binding:vsphere [Feature:vsphere] should bind volume with claim for given label" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] new files should be created with FSGroup ownership when container is root" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support two pods which share the same volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce policy based on PodSelector or NamespaceSelector [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes NFS when invoking the Recycle reclaim policy should test that a PV becomes Available and is clean after the PVC is deleted." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Attach Verify [Feature:vsphere][Serial][Disruptive] verify volume remains attached after master kubelet restart" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-cloud-provider-gcp] kube-proxy migration [Feature:KubeProxyDaemonSetMigration] Upgrade kube-proxy from static pods to a DaemonSet should maintain a functioning cluster [Feature:KubeProxyDaemonSetUpgrade]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Pod Disks schedule a pod w/ RW PD(s) mounted to 1 or more containers, write to PD, verify content, delete pod, and repeat in rapid succession [Slow] using 4 containers and 1 PDs" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSIStorageCapacity [Feature:CSIStorageCapacity] CSIStorageCapacity unused" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should serve multiport endpoints from pods  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable-stress[Feature:VolumeSnapshotDataSource] should support snapshotting of many volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-cloud-provider-gcp] Ports Security Check [Feature:KubeletSecurity] should not have port 4194 open on its all public IP addresses" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Secrets should patch a secret [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Services should function for pod-Service: sctp [Feature:SCTPConnectivity]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link] Set fsGroup for local volume should set fsGroup for one pod [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] IngressClass [Feature:Ingress] should set default value on new IngressClass [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Sysctls [LinuxOnly] [NodeFeature:Sysctls] should not launch unsafe, but not explicitly enabled sysctls on the node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce policy to allow traffic only from a different namespace, based on NamespaceSelector [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] NFSPersistentVolumes[Disruptive][Flaky] when kube-controller-manager restarts should delete a bound PVC from a clientPod, restart the kube-control-manager, and ensure the kube-controller-manager does not crash" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Service endpoints latency should not be very high  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates that taints-tolerations is respected if matching" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DaemonRestart [Disruptive] Scheduler should continue assigning pods to nodes across restart" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] SCTP [Feature:SCTP] [LinuxOnly] should allow creating a basic SCTP service with pod and endpoints" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] vsphere statefulset [Feature:vsphere] vsphere statefulset testing" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:Ingress] should support multiple TLS certs" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] RuntimeClass should reject a Pod requesting a RuntimeClass with conflicting node selector" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should support a &#39;default-deny-all&#39; policy [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] nonexistent volume subPath should have the correct mode and owner using FSGroup" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should work after restarting kube-proxy [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] [Feature:Windows] GMSA Full [Slow] GMSA support works end to end" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Pods should be updated [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI attach test using mock driver should require VolumeAttach for drivers with attachment" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should have session affinity work for LoadBalancer service with ESIPP off [Slow] [DisabledForLargeClusters] [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Services should update endpoints: udp" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Storage Policy Based Volume Provisioning [Feature:vsphere] verify if a SPBM policy is not honored on a non-compatible datastore for dynamically provisioned pvc using storageclass" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] ClusterDns [Feature:Example] should create pod that uses dns" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Events should delete a collection of events [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Character Device [Slow] Should be able to mount character device &#39;achardev&#39; successfully when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV, based on the allowed zones specified in storage class when the datastore under the zone is present in another datacenter" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Pods Extended [k8s.io] Pod Container lifecycle should not create extra sandbox if all containers are done" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] ReplicaSet should serve a basic image on each replica with a public image  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] [Feature:NodeAuthorizer] A node shouldn&#39;t be able to delete another node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Network should resolve connection reset issue #74839 [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes [Feature:vsphere][Feature:ReclaimPolicy] [sig-storage] persistentvolumereclaim:vsphere [Feature:vsphere] should retain persistent volume when reclaimPolicy set to retain when associated claim is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Disk Format [Feature:vsphere] verify disk format type - eagerzeroedthick is honored for dynamically provisioned pv using storageclass" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PVC Protection Verify that PVC in active use by a pod is not removed immediately" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Simple pod should support exec using resource/name" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: gce-localssd-scsi-fs] [Serial] One pod requesting one prebound PVC should be able to mount volume and write from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Sysctls [LinuxOnly] [NodeFeature:Sysctls] should support unsafe sysctls which are actually whitelisted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PV Protection Verify that PV bound to a PVC is not removed immediately" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Cluster size autoscaler scalability [Slow] should scale up twice [Feature:ClusterAutoscalerScalability2]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory &#39;adir&#39; when HostPathType is HostPathSocket" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] AppArmor load AppArmor profiles can disable an AppArmor profile, using unconfined" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStackAlphaFeature] [LinuxOnly] should have ipv4 and ipv6 node podCIDRs" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Downward API should provide host IP and pod IP as an env var if pod uses host network [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] ServiceAccounts should ensure a single API token exists" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify PVC creation with incompatible zone along with compatible storagePolicy and datastore combination specified in storage class fails" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Placement [Feature:vsphere] should create and delete pod with the same volume source attach/detach to different worker nodes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should disable node pool autoscaling [Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: gce-localssd-scsi-fs] [Serial] Set fsGroup for local volume should set fsGroup for one pod [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] shouldn&#39;t increase cluster size if pending pod is too large [Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV, based on multiple zones specified in the storage class. (No shared datastores exist among both zones)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  StatefulSet with pod affinity [Slow] should use volumes spread across nodes when pod management is parallel and pod has anti-affinity" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: tmpfs] One pod requesting one prebound PVC should be able to mount volume and write from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:NEG] should create NEGs for all ports with the Ingress annotation, and NEGs for the standalone annotation otherwise" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl apply apply set/view last-applied" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should find a service from listing all namespaces [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] Multi-AZ Cluster Volumes [sig-storage] should only be allowed to provision PDs in zones where nodes exist" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI NodeStage error cases [Slow] should call NodeUnstage after NodeStage success" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Storage Policy Based Volume Provisioning [Feature:vsphere] verify clean up of stale dummy VM for dynamically provisioned pvc using SPBM policy" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] PreStop graceful pod terminated should wait until preStop hook completes the process [Flaky]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] files with FSGroup ownership should support (root,0644,tmpfs)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Storage Policy Based Volume Provisioning [Feature:vsphere] verify VSAN storage capability with valid diskStripes and objectSpaceReservation values and a VSAN datastore is honored for dynamically provisioned pvc using storageclass" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Pods should have their auto-restart back-off timer reset on image update [Slow][NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Mount propagation should propagate mounts to the host" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Etcd failure [Disruptive] should recover from network partition with master" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : secret" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Hostname of Pod [Feature:SetHostnameAsFQDN] a pod with subdomain field has FQDN, hostname is shortname [Feature:SetHostnameAsFQDN]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Servers with support for Table transformation should return generic metadata details across all namespaces for nodes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:NEG] should be able to switch between IG and NEG modes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device &#39;ablkdev&#39; when HostPathType is HostPathSocket" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should implement service.kubernetes.io/service-proxy-name" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Provisioning On Clustered Datastore [Feature:vsphere] verify dynamic provision with spbm policy on clustered datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates that NodeAffinity is respected if not matching" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PVC Protection Verify that scheduling of a pod that uses PVC that is being deleted fails and the pod becomes Unschedulable" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir] One pod requesting one prebound PVC should be able to mount volume and read from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should be able to scale down when rescheduling a pod is required and pdb allows for it[Feature:ClusterSizeAutoscalingScaleDown]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking should recreate its iptables rules if they are deleted [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl create quota should reject quota with invalid scopes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] GenericPersistentVolume[Disruptive] When kubelet restarts Should test that a volume mounted to a pod that is force deleted while the kubelet is down unmounts when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] [Serial] Volume metrics should create prometheus metrics for volume provisioning and attach/detach" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Downward API should provide container&#39;s limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Generic Ephemeral-volume (default fs) [Feature:GenericEphemeralVolume] (immediate-binding)] ephemeral should support multiple inline ephemeral volumes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [Feature:Example] [k8s.io] Liveness liveness pods should be automatically restarted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] Services should be able to create a functioning NodePort service for Windows" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Variable Expansion should verify that a failing subpath expansion can be modified during the lifecycle of a container [sig-storage][Slow] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Pod Disks schedule pods each with a PD, delete pod and verify detach [Slow] for RW PD with pod delete grace period of &#34;default (30s)&#34;" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: tmpfs] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Container Runtime blackbox test when running a container with a new image should not be able to pull from private registry without secret [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Hostname of Pod [Feature:SetHostnameAsFQDN] a pod with subdomain field has FQDN, when setHostnameAsFQDN is set to true, the FQDN is set as hostname [Feature:SetHostnameAsFQDN]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify PVC creation with invalid zone specified in storage class fails" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-cloud-provider] [Feature:CloudProvider][Disruptive] Nodes should be deleted on API server if it doesn&#39;t exist in the cloud provider" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] GCP Volumes NFSv4 should be mountable for NFSv4" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI workload information using mock driver should be passed when podInfoOnMount=true" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable-stress[Feature:VolumeSnapshotDataSource] should support snapshotting of many volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Garbage collector should support cascading deletion of custom resources" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI volume limit information using mock driver should report attach limit when limit is bigger than 0 [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Deployment iterative rollouts should eventually progress" classname="Kubernetes e2e suite" time="333.76885382">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:121&#xA;May 12 19:42:08.366: Unexpected error:&#xA;    &lt;*errors.errorString | 0xc001cc1660&gt;: {&#xA;        s: &#34;error waiting for deployment \&#34;webserver\&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:\&#34;Available\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;MinimumReplicasUnavailable\&#34;, Message:\&#34;Deployment does not have minimum availability.\&#34;}, v1.DeploymentCondition{Type:\&#34;Progressing\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;ProgressDeadlineExceeded\&#34;, Message:\&#34;ReplicaSet \\\&#34;webserver-5f9c9b577c\\\&#34; has timed out progressing.\&#34;}}, CollisionCount:(*int32)(nil)}&#34;,&#xA;    }&#xA;    error waiting for deployment &#34;webserver&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;occurred&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:648</failure>
          <system-out>[BeforeEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174&#xA;STEP: Creating a kubernetes client&#xA;May 12 19:36:35.280: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename deployment&#xA;May 12 19:36:35.298: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled&#xA;May 12 19:36:35.300: INFO: Error creating dryrun pod; assuming PodSecurityPolicy is disabled: admission webhook &#34;cmk.intel.com&#34; does not support dry run&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[BeforeEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:78&#xA;[It] iterative rollouts should eventually progress&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:121&#xA;May 12 19:36:35.302: INFO: Creating deployment &#34;webserver&#34;&#xA;May 12 19:36:35.306: INFO: 00: resuming deployment &#34;webserver&#34;&#xA;May 12 19:36:35.308: INFO: 00: scaling down&#xA;May 12 19:36:35.311: INFO: Updating deployment webserver&#xA;May 12 19:36:35.892: INFO: 01: scaling deployment &#34;webserver&#34;&#xA;May 12 19:36:35.898: INFO: Updating deployment webserver&#xA;May 12 19:36:36.731: INFO: 02: scaling deployment &#34;webserver&#34;&#xA;May 12 19:36:36.744: INFO: Updating deployment webserver&#xA;May 12 19:36:36.744: INFO: 03: triggering a new rollout for deployment &#34;webserver&#34;&#xA;May 12 19:36:36.750: INFO: Updating deployment webserver&#xA;May 12 19:36:37.574: INFO: 04: arbitrarily deleting one or more deployment pods for deployment &#34;webserver&#34;&#xA;May 12 19:36:37.578: INFO: 04: deleting deployment pod &#34;webserver-5f9c9b577c-q9rdk&#34;&#xA;May 12 19:36:37.586: INFO: 04: deleting deployment pod &#34;webserver-5f9c9b577c-vl7j4&#34;&#xA;May 12 19:36:37.592: INFO: 04: deleting deployment pod &#34;webserver-dd94f59b7-gf9ss&#34;&#xA;May 12 19:36:37.597: INFO: 04: deleting deployment pod &#34;webserver-dd94f59b7-krdg8&#34;&#xA;May 12 19:36:38.019: INFO: 05: resuming deployment &#34;webserver&#34;&#xA;May 12 19:36:38.024: INFO: Updating deployment webserver&#xA;May 12 19:36:39.449: INFO: 06: resuming deployment &#34;webserver&#34;&#xA;May 12 19:36:39.452: INFO: 06: scaling down&#xA;May 12 19:36:39.457: INFO: Updating deployment webserver&#xA;May 12 19:36:40.401: INFO: 07: resuming deployment &#34;webserver&#34;&#xA;May 12 19:36:40.404: INFO: 07: scaling down&#xA;May 12 19:36:40.408: INFO: Updating deployment webserver&#xA;May 12 19:36:42.910: INFO: 08: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 12 19:36:42.918: INFO: Updating deployment webserver&#xA;May 12 19:36:43.788: INFO: 09: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 12 19:36:43.795: INFO: Updating deployment webserver&#xA;May 12 19:36:44.752: INFO: 10: arbitrarily deleting one or more deployment pods for deployment &#34;webserver&#34;&#xA;May 12 19:36:44.755: INFO: 10: deleting deployment pod &#34;webserver-dd94f59b7-2lpbc&#34;&#xA;May 12 19:36:44.764: INFO: 10: deleting deployment pod &#34;webserver-dd94f59b7-m7bp9&#34;&#xA;May 12 19:36:47.754: INFO: 11: scaling deployment &#34;webserver&#34;&#xA;May 12 19:36:47.757: INFO: 11: scaling up&#xA;May 12 19:36:47.762: INFO: Updating deployment webserver&#xA;May 12 19:36:49.628: INFO: 12: arbitrarily deleting one or more deployment pods for deployment &#34;webserver&#34;&#xA;May 12 19:36:49.632: INFO: 12: deleting deployment pod &#34;webserver-dd94f59b7-r4hw8&#34;&#xA;May 12 19:36:53.006: INFO: 13: scaling deployment &#34;webserver&#34;&#xA;May 12 19:36:53.013: INFO: Updating deployment webserver&#xA;May 12 19:36:58.421: INFO: 14: scaling deployment &#34;webserver&#34;&#xA;May 12 19:36:58.426: INFO: Updating deployment webserver&#xA;May 12 19:37:01.234: INFO: 15: scaling deployment &#34;webserver&#34;&#xA;May 12 19:37:01.236: INFO: 15: scaling down&#xA;May 12 19:37:01.240: INFO: Updating deployment webserver&#xA;May 12 19:37:01.240: INFO: 16: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 12 19:37:01.246: INFO: Updating deployment webserver&#xA;May 12 19:37:01.247: INFO: 17: triggering a new rollout for deployment &#34;webserver&#34;&#xA;May 12 19:37:01.248: INFO: 17: scaling up&#xA;May 12 19:37:01.252: INFO: Updating deployment webserver&#xA;May 12 19:37:06.323: INFO: 18: arbitrarily deleting one or more deployment pods for deployment &#34;webserver&#34;&#xA;May 12 19:37:06.326: INFO: 18: deleting deployment pod &#34;webserver-6699947579-jqfsm&#34;&#xA;May 12 19:37:06.333: INFO: 18: deleting deployment pod &#34;webserver-dd94f59b7-wclqf&#34;&#xA;May 12 19:37:06.340: INFO: 19: resuming deployment &#34;webserver&#34;&#xA;May 12 19:37:06.343: INFO: 19: scaling up&#xA;May 12 19:37:08.353: INFO: Updating deployment webserver&#xA;May 12 19:37:08.355: INFO: Waiting for deployment &#34;webserver&#34; to be observed by the controller&#xA;May 12 19:37:08.357: INFO: Waiting for deployment &#34;webserver&#34; status&#xA;May 12 19:37:08.359: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:10.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:12.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:14.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:16.367: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:18.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:20.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:22.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:24.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:26.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:28.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:30.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:32.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:34.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:36.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445026, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:38.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:40.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:42.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:44.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:46.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:48.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:50.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:52.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:54.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:56.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:37:58.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:00.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:02.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:04.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:06.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:08.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:10.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:12.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:14.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:16.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:18.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:20.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:22.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:24.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:26.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:28.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:30.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:32.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:34.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:36.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:38.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:40.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:42.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:44.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:46.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:48.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:50.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:52.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:54.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:56.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:38:58.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:00.367: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:02.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:04.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:06.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:08.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:10.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:12.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:14.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:16.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:18.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:20.367: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:22.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:24.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:26.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:28.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:30.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:32.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:34.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:36.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:38.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:40.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:42.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:44.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:46.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:48.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:50.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:52.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:54.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:56.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:39:58.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:00.366: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:02.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:04.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:06.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:08.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:10.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:12.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:14.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:16.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:18.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:20.367: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:22.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:24.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:26.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:28.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:30.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:32.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:34.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:36.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:38.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:40.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:42.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:44.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:46.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:48.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:50.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:52.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:54.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:56.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:40:58.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:00.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:02.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:04.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:06.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:08.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:10.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:12.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:14.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:16.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:18.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:20.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:22.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:24.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:26.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:28.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:30.364: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:32.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:34.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:36.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:38.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:40.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:42.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:44.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:46.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:48.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:50.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:52.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:54.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:56.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:41:58.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:42:00.362: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:42:02.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:42:04.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:42:06.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:42:08.363: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:42:08.365: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 12 19:42:08.366: FAIL: Unexpected error:&#xA;    &lt;*errors.errorString | 0xc001cc1660&gt;: {&#xA;        s: &#34;error waiting for deployment \&#34;webserver\&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:\&#34;Available\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;MinimumReplicasUnavailable\&#34;, Message:\&#34;Deployment does not have minimum availability.\&#34;}, v1.DeploymentCondition{Type:\&#34;Progressing\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;ProgressDeadlineExceeded\&#34;, Message:\&#34;ReplicaSet \\\&#34;webserver-5f9c9b577c\\\&#34; has timed out progressing.\&#34;}}, CollisionCount:(*int32)(nil)}&#34;,&#xA;    }&#xA;    error waiting for deployment &#34;webserver&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:14, Replicas:5, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:5, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756444995, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756445057, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-5f9c9b577c\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;occurred&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/apps.testIterativeDeployments(0xc000a1ba20)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:648 +0x184f&#xA;k8s.io/kubernetes/test/e2e/apps.glob..func4.8()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:122 +0x2a&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc000601c80)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc000601c80)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc000601c80, 0x4de37a0)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;[AfterEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72&#xA;May 12 19:42:08.372: INFO: Deployment &#34;webserver&#34;:&#xA;&amp;Deployment{ObjectMeta:{webserver  deployment-4463 /apis/apps/v1/namespaces/deployment-4463/deployments/webserver f38ecca0-ed09-4904-8d7f-140f8336f136 69271 14 2021-05-12 19:36:35 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd] map[deployment.kubernetes.io/revision:6] [] []  []},Spec:DeploymentSpec{Replicas:*4,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [{A 3 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc000505ee8 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&amp;RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*2,Paused:false,ProgressDeadlineSeconds:*30,},Status:DeploymentStatus{ObservedGeneration:14,Replicas:5,UpdatedReplicas:2,AvailableReplicas:0,UnavailableReplicas:5,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2021-05-12 19:36:35 +0000 UTC,LastTransitionTime:2021-05-12 19:36:35 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:False,Reason:ProgressDeadlineExceeded,Message:ReplicaSet &#34;webserver-5f9c9b577c&#34; has timed out progressing.,LastUpdateTime:2021-05-12 19:37:37 +0000 UTC,LastTransitionTime:2021-05-12 19:37:37 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},}&#xA;&#xA;May 12 19:42:08.375: INFO: New ReplicaSet &#34;webserver-5f9c9b577c&#34; of Deployment &#34;webserver&#34;:&#xA;&amp;ReplicaSet{ObjectMeta:{webserver-5f9c9b577c  deployment-4463 /apis/apps/v1/namespaces/deployment-4463/replicasets/webserver-5f9c9b577c d50571d2-28a6-430c-afda-50c507de3a77 68573 5 2021-05-12 19:36:36 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5f9c9b577c] map[deployment.kubernetes.io/desired-replicas:4 deployment.kubernetes.io/max-replicas:5 deployment.kubernetes.io/revision:6 deployment.kubernetes.io/revision-history:2,4] [{apps/v1 Deployment webserver f38ecca0-ed09-4904-8d7f-140f8336f136 0xc00060b4b0 0xc00060b4b1}] []  [{kube-controller-manager Update apps/v1 2021-05-12 19:36:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:deployment.kubernetes.io/desired-replicas&#34;:{},&#34;f:deployment.kubernetes.io/max-replicas&#34;:{},&#34;f:deployment.kubernetes.io/revision&#34;:{},&#34;f:deployment.kubernetes.io/revision-history&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;f38ecca0-ed09-4904-8d7f-140f8336f136\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:replicas&#34;:{},&#34;f:selector&#34;:{&#34;f:matchLabels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:template&#34;:{&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:env&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;name\&#34;:\&#34;A\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:value&#34;:{}}},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}},&#34;f:status&#34;:{&#34;f:fullyLabeledReplicas&#34;:{},&#34;f:observedGeneration&#34;:{},&#34;f:replicas&#34;:{}}}}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 5f9c9b577c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5f9c9b577c] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [{A 3 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00060b558 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:2,FullyLabeledReplicas:2,ObservedGeneration:5,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}&#xA;May 12 19:42:08.375: INFO: All old ReplicaSets of Deployment &#34;webserver&#34;:&#xA;May 12 19:42:08.376: INFO: &amp;ReplicaSet{ObjectMeta:{webserver-dd94f59b7  deployment-4463 /apis/apps/v1/namespaces/deployment-4463/replicasets/webserver-dd94f59b7 d0de0868-56ba-4793-b0f4-269548ebe2bd 68921 5 2021-05-12 19:36:35 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:dd94f59b7] map[deployment.kubernetes.io/desired-replicas:4 deployment.kubernetes.io/max-replicas:5 deployment.kubernetes.io/revision:3 deployment.kubernetes.io/revision-history:1] [{apps/v1 Deployment webserver f38ecca0-ed09-4904-8d7f-140f8336f136 0xc00060b6e0 0xc00060b6e1}] []  [{kube-controller-manager Update apps/v1 2021-05-12 19:36:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:deployment.kubernetes.io/desired-replicas&#34;:{},&#34;f:deployment.kubernetes.io/max-replicas&#34;:{},&#34;f:deployment.kubernetes.io/revision&#34;:{},&#34;f:deployment.kubernetes.io/revision-history&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;f38ecca0-ed09-4904-8d7f-140f8336f136\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:replicas&#34;:{},&#34;f:selector&#34;:{&#34;f:matchLabels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:template&#34;:{&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}},&#34;f:status&#34;:{&#34;f:fullyLabeledReplicas&#34;:{},&#34;f:observedGeneration&#34;:{},&#34;f:replicas&#34;:{}}}}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: dd94f59b7,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:dd94f59b7] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00060b748 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:2,FullyLabeledReplicas:2,ObservedGeneration:5,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}&#xA;May 12 19:42:08.376: INFO: &amp;ReplicaSet{ObjectMeta:{webserver-6699947579  deployment-4463 /apis/apps/v1/namespaces/deployment-4463/replicasets/webserver-6699947579 b2e243f1-8cda-44dd-b93c-7eae3c075635 68913 1 2021-05-12 19:37:01 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:6699947579] map[deployment.kubernetes.io/desired-replicas:4 deployment.kubernetes.io/max-replicas:5 deployment.kubernetes.io/revision:5] [{apps/v1 Deployment webserver f38ecca0-ed09-4904-8d7f-140f8336f136 0xc00060b5c0 0xc00060b5c1}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 6699947579,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:6699947579] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [{A 3 nil} {A 17 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00060b678 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}&#xA;May 12 19:42:08.380: INFO: Pod &#34;webserver-5f9c9b577c-5bnpn&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-5f9c9b577c-5bnpn webserver-5f9c9b577c- deployment-4463 /api/v1/namespaces/deployment-4463/pods/webserver-5f9c9b577c-5bnpn 073c25a0-7979-4adb-86f7-590be82139fa 70436 0 2021-05-12 19:36:37 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5f9c9b577c] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.250&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;de:52:66:0b:33:e4&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.250&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;de:52:66:0b:33:e4&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-5f9c9b577c d50571d2-28a6-430c-afda-50c507de3a77 0xc0001a409f 0xc0001a4180}] []  [{kube-controller-manager Update v1 2021-05-12 19:36:37 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:generateName&#34;:{},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;d50571d2-28a6-430c-afda-50c507de3a77\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:env&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;name\&#34;:\&#34;A\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:value&#34;:{}}},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:enableServiceLinks&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}} {multus Update v1 2021-05-12 19:36:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:k8s.v1.cni.cncf.io/network-status&#34;:{},&#34;f:k8s.v1.cni.cncf.io/networks-status&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:37:00 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;ContainersReady\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Initialized\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:containerStatuses&#34;:{},&#34;f:hostIP&#34;:{},&#34;f:podIP&#34;:{},&#34;f:podIPs&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;ip\&#34;:\&#34;10.244.3.250\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:ip&#34;:{}}},&#34;f:startTime&#34;:{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-976lc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-976lc,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A,Value:3,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-976lc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:37 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:37 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:37 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:37 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.207,PodIP:10.244.3.250,StartTime:2021-05-12 19:36:37 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.3.250,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;May 12 19:42:08.380: INFO: Pod &#34;webserver-5f9c9b577c-zk25n&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-5f9c9b577c-zk25n webserver-5f9c9b577c- deployment-4463 /api/v1/namespaces/deployment-4463/pods/webserver-5f9c9b577c-zk25n ffb7e8a7-73fe-4399-9fc3-7b449f2a9184 70450 0 2021-05-12 19:36:36 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5f9c9b577c] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.71&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;0a:50:fc:36:66:f1&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.71&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;0a:50:fc:36:66:f1&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-5f9c9b577c d50571d2-28a6-430c-afda-50c507de3a77 0xc0001a471f 0xc0001a4740}] []  [{kube-controller-manager Update v1 2021-05-12 19:36:36 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:generateName&#34;:{},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;d50571d2-28a6-430c-afda-50c507de3a77\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:env&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;name\&#34;:\&#34;A\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:value&#34;:{}}},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:enableServiceLinks&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}} {kubelet Update v1 2021-05-12 19:36:56 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;ContainersReady\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Initialized\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:containerStatuses&#34;:{},&#34;f:hostIP&#34;:{},&#34;f:podIP&#34;:{},&#34;f:podIPs&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;ip\&#34;:\&#34;10.244.4.71\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:ip&#34;:{}}},&#34;f:startTime&#34;:{}}}} {multus Update v1 2021-05-12 19:36:56 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:k8s.v1.cni.cncf.io/network-status&#34;:{},&#34;f:k8s.v1.cni.cncf.io/networks-status&#34;:{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-976lc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-976lc,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A,Value:3,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-976lc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:36 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:36 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:36 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.208,PodIP:10.244.4.71,StartTime:2021-05-12 19:36:36 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.4.71,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;May 12 19:42:08.380: INFO: Pod &#34;webserver-6699947579-4rbmr&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-6699947579-4rbmr webserver-6699947579- deployment-4463 /api/v1/namespaces/deployment-4463/pods/webserver-6699947579-4rbmr 2ef6af25-cc62-481c-ac58-f3a9cfbe7334 70473 0 2021-05-12 19:37:06 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:6699947579] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.78&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;86:dd:d3:7b:06:8b&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.78&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;86:dd:d3:7b:06:8b&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-6699947579 b2e243f1-8cda-44dd-b93c-7eae3c075635 0xc0001a4ddf 0xc0001a4e00}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-976lc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-976lc,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A,Value:3,ValueFrom:nil,},EnvVar{Name:A,Value:17,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-976lc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.208,PodIP:10.244.4.78,StartTime:2021-05-12 19:37:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.4.78,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;May 12 19:42:08.381: INFO: Pod &#34;webserver-dd94f59b7-6pgrf&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-dd94f59b7-6pgrf webserver-dd94f59b7- deployment-4463 /api/v1/namespaces/deployment-4463/pods/webserver-dd94f59b7-6pgrf 49b45d3c-09d3-4d45-bc20-c9d3852f3760 70514 0 2021-05-12 19:37:06 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:dd94f59b7] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.79&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;de:0b:aa:c2:42:2b&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.79&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;de:0b:aa:c2:42:2b&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-dd94f59b7 d0de0868-56ba-4793-b0f4-269548ebe2bd 0xc0001a528f 0xc0001a52b0}] []  [{kube-controller-manager Update v1 2021-05-12 19:37:06 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:generateName&#34;:{},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;d0de0868-56ba-4793-b0f4-269548ebe2bd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:enableServiceLinks&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}} {kubelet Update v1 2021-05-12 19:37:18 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;ContainersReady\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Initialized\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:containerStatuses&#34;:{},&#34;f:hostIP&#34;:{},&#34;f:podIP&#34;:{},&#34;f:podIPs&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;ip\&#34;:\&#34;10.244.4.79\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:ip&#34;:{}}},&#34;f:startTime&#34;:{}}}} {multus Update v1 2021-05-12 19:37:18 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:k8s.v1.cni.cncf.io/network-status&#34;:{},&#34;f:k8s.v1.cni.cncf.io/networks-status&#34;:{}}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-976lc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-976lc,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-976lc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:37:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.208,PodIP:10.244.4.79,StartTime:2021-05-12 19:37:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.4.79,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;May 12 19:42:08.381: INFO: Pod &#34;webserver-dd94f59b7-dll2m&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-dd94f59b7-dll2m webserver-dd94f59b7- deployment-4463 /api/v1/namespaces/deployment-4463/pods/webserver-dd94f59b7-dll2m abde5232-22b4-4659-a777-62576bf413aa 70492 0 2021-05-12 19:36:49 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:dd94f59b7] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.248&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;e6:32:42:7c:45:d4&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.248&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;e6:32:42:7c:45:d4&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-dd94f59b7 d0de0868-56ba-4793-b0f4-269548ebe2bd 0xc0001a5d1f 0xc0001a5d30}] []  [{kube-controller-manager Update v1 2021-05-12 19:36:49 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:generateName&#34;:{},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;d0de0868-56ba-4793-b0f4-269548ebe2bd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:enableServiceLinks&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}} {multus Update v1 2021-05-12 19:36:52 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:k8s.v1.cni.cncf.io/network-status&#34;:{},&#34;f:k8s.v1.cni.cncf.io/networks-status&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:36:59 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;ContainersReady\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Initialized\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:containerStatuses&#34;:{},&#34;f:hostIP&#34;:{},&#34;f:podIP&#34;:{},&#34;f:podIPs&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;ip\&#34;:\&#34;10.244.3.248\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:ip&#34;:{}}},&#34;f:startTime&#34;:{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-976lc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-976lc,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-976lc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:49 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:49 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:49 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-12 19:36:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.207,PodIP:10.244.3.248,StartTime:2021-05-12 19:36:49 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.3.248,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;[AfterEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175&#xA;STEP: Collecting events from namespace &#34;deployment-4463&#34;.&#xA;STEP: Found 125 events.&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: { } Scheduled: Successfully assigned deployment-4463/webserver-5f9c9b577c-5bnpn to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5f9c9b577c-q9rdk: { } Scheduled: Successfully assigned deployment-4463/webserver-5f9c9b577c-q9rdk to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5f9c9b577c-r2ztq: { } Scheduled: Successfully assigned deployment-4463/webserver-5f9c9b577c-r2ztq to node2&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5f9c9b577c-vl7j4: { } Scheduled: Successfully assigned deployment-4463/webserver-5f9c9b577c-vl7j4 to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5f9c9b577c-wjqqs: { } Scheduled: Successfully assigned deployment-4463/webserver-5f9c9b577c-wjqqs to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5f9c9b577c-zk25n: { } Scheduled: Successfully assigned deployment-4463/webserver-5f9c9b577c-zk25n to node2&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-6699947579-4rbmr: { } Scheduled: Successfully assigned deployment-4463/webserver-6699947579-4rbmr to node2&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-6699947579-jqfsm: { } Scheduled: Successfully assigned deployment-4463/webserver-6699947579-jqfsm to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-2lpbc: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-2lpbc to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-2s9sc: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-2s9sc to node2&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-4s82c: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-4s82c to node2&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-5f2sj: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-5f2sj to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-6pgrf: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-6pgrf to node2&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-dll2m: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-dll2m to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-gf9ss: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-gf9ss to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-krdg8: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-krdg8 to node1&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-m7bp9: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-m7bp9 to node2&#xA;May 12 19:42:08.387: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-ntg6v: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-ntg6v to node1&#xA;May 12 19:42:08.388: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-r4hw8: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-r4hw8 to node1&#xA;May 12 19:42:08.388: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-wclqf: { } Scheduled: Successfully assigned deployment-4463/webserver-dd94f59b7-wclqf to node2&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-dd94f59b7 to 6&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-dd94f59b7 to 5&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-m7bp9&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-2s9sc&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-krdg8&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-4s82c&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-5f2sj&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:35 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-gf9ss&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:36 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-5f9c9b577c to 2&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:36 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-dd94f59b7 to 4&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:36 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-5f9c9b577c to 3&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:36 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5f9c9b577c-vl7j4&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:36 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5f9c9b577c-zk25n&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:36 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-5f2sj&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:36 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-2s9sc&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:37 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5f9c9b577c-5bnpn&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:37 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5f9c9b577c-q9rdk&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:37 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-2lpbc&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:38 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5f9c9b577c-r2ztq&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:38 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-ntg6v&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:39 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-5f9c9b577c to 2&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:39 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-dd94f59b7 to 3&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:39 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-5f9c9b577c-r2ztq&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:39 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-4s82c&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:40 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-dd94f59b7 to 2&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:40 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-ntg6v&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:42 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 1&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:42 +0000 UTC - event for webserver-dd94f59b7-4s82c: {kubelet node2} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to start sandbox container for pod &#34;webserver-dd94f59b7-4s82c&#34;: Error response from daemon: OCI runtime create failed: container_linux.go:349: starting container process caused &#34;process_linux.go:315: copying bootstrap data to pipe caused \&#34;write init-p: broken pipe\&#34;&#34;: unknown&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:43 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 2&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:44 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-r4hw8&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:44 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: (combined from similar events): Created pod: webserver-dd94f59b7-6pgrf&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:45 +0000 UTC - event for webserver-dd94f59b7-m7bp9: {multus } AddedInterface: Add eth0 [10.244.4.64/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:46 +0000 UTC - event for webserver-dd94f59b7-krdg8: {kubelet node1} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to start sandbox container for pod &#34;webserver-dd94f59b7-krdg8&#34;: Error response from daemon: OCI runtime create failed: container_linux.go:349: starting container process caused &#34;process_linux.go:319: getting the final child&#39;s pid from pipe caused \&#34;EOF\&#34;&#34;: unknown&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:46 +0000 UTC - event for webserver-dd94f59b7-m7bp9: {kubelet node2} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to set up sandbox container &#34;4565b44423afd55f70b36479ab4e425ab0f703f7ce4090f4af14ccc01b927c74&#34; network for pod &#34;webserver-dd94f59b7-m7bp9&#34;: networkPlugin cni failed to set up pod &#34;webserver-dd94f59b7-m7bp9_deployment-4463&#34; network: Multus: [deployment-4463/webserver-dd94f59b7-m7bp9]: error setting the networks status: SetNetworkStatus: failed to query the pod webserver-dd94f59b7-m7bp9 in out of cluster comm: pods &#34;webserver-dd94f59b7-m7bp9&#34; not found&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:47 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5f9c9b577c-wjqqs&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:49 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:49 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {multus } AddedInterface: Add eth0 [10.244.3.246/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:49 +0000 UTC - event for webserver-dd94f59b7-r4hw8: {multus } AddedInterface: Add eth0 [10.244.3.245/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:50 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {multus } AddedInterface: Add eth0 [10.244.4.65/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:50 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:50 +0000 UTC - event for webserver-dd94f59b7-r4hw8: {kubelet node1} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to set up sandbox container &#34;662f291079e9fa82654140ed524d7c5df3ceb04d252bd27b5e344b63a9ace7da&#34; network for pod &#34;webserver-dd94f59b7-r4hw8&#34;: networkPlugin cni failed to set up pod &#34;webserver-dd94f59b7-r4hw8_deployment-4463&#34; network: Multus: [deployment-4463/webserver-dd94f59b7-r4hw8]: error setting the networks status: SetNetworkStatus: failed to query the pod webserver-dd94f59b7-r4hw8 in out of cluster comm: pods &#34;webserver-dd94f59b7-r4hw8&#34; not found&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:50 +0000 UTC - event for webserver-dd94f59b7-wclqf: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:50 +0000 UTC - event for webserver-dd94f59b7-wclqf: {multus } AddedInterface: Add eth0 [10.244.4.68/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:51 +0000 UTC - event for webserver-5f9c9b577c-wjqqs: {multus } AddedInterface: Add eth0 [10.244.3.247/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:51 +0000 UTC - event for webserver-5f9c9b577c-wjqqs: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:51 +0000 UTC - event for webserver-dd94f59b7-wclqf: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:51 +0000 UTC - event for webserver-dd94f59b7-wclqf: {kubelet node2} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:51 +0000 UTC - event for webserver-dd94f59b7-wclqf: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:52 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:52 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:52 +0000 UTC - event for webserver-dd94f59b7-dll2m: {multus } AddedInterface: Add eth0 [10.244.3.248/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:52 +0000 UTC - event for webserver-dd94f59b7-dll2m: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:53 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {kubelet node2} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:53 +0000 UTC - event for webserver-dd94f59b7-wclqf: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:53 +0000 UTC - event for webserver-dd94f59b7-wclqf: {multus } AddedInterface: Add eth0 [10.244.4.69/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:53 +0000 UTC - event for webserver-dd94f59b7-wclqf: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:56 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:56 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:56 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:56 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:56 +0000 UTC - event for webserver-5f9c9b577c-zk25n: {multus } AddedInterface: Add eth0 [10.244.4.71/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:57 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {kubelet node1} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:57 +0000 UTC - event for webserver-5f9c9b577c-wjqqs: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:57 +0000 UTC - event for webserver-5f9c9b577c-wjqqs: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:57 +0000 UTC - event for webserver-5f9c9b577c-wjqqs: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:57 +0000 UTC - event for webserver-5f9c9b577c-wjqqs: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:58 +0000 UTC - event for webserver-dd94f59b7-dll2m: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:58 +0000 UTC - event for webserver-dd94f59b7-dll2m: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:59 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:59 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:59 +0000 UTC - event for webserver-5f9c9b577c-5bnpn: {multus } AddedInterface: Add eth0 [10.244.3.250/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:59 +0000 UTC - event for webserver-dd94f59b7-dll2m: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:36:59 +0000 UTC - event for webserver-dd94f59b7-dll2m: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:01 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-6699947579 to 1&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:01 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 4&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:01 +0000 UTC - event for webserver-5f9c9b577c: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-5f9c9b577c-wjqqs&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:01 +0000 UTC - event for webserver-6699947579: {replicaset-controller } SuccessfulCreate: Created pod: webserver-6699947579-jqfsm&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:06 +0000 UTC - event for webserver-6699947579: {replicaset-controller } SuccessfulCreate: Created pod: webserver-6699947579-4rbmr&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:08 +0000 UTC - event for webserver-6699947579-jqfsm: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:10 +0000 UTC - event for webserver-6699947579-jqfsm: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:10 +0000 UTC - event for webserver-6699947579-jqfsm: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:12 +0000 UTC - event for webserver-6699947579-4rbmr: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:12 +0000 UTC - event for webserver-6699947579-4rbmr: {multus } AddedInterface: Add eth0 [10.244.4.76/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:12 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:12 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {multus } AddedInterface: Add eth0 [10.244.4.77/24]&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:14 +0000 UTC - event for webserver-6699947579-4rbmr: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:14 +0000 UTC - event for webserver-6699947579-4rbmr: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:15 +0000 UTC - event for webserver-6699947579-4rbmr: {kubelet node2} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:15 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:42:08.388: INFO: At 2021-05-12 19:37:15 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:37:15 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {kubelet node2} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:37:17 +0000 UTC - event for webserver-6699947579-4rbmr: {multus } AddedInterface: Add eth0 [10.244.4.78/24]&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:37:17 +0000 UTC - event for webserver-6699947579-4rbmr: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:37:17 +0000 UTC - event for webserver-6699947579-4rbmr: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:37:18 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {multus } AddedInterface: Add eth0 [10.244.4.79/24]&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:37:18 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:37:18 +0000 UTC - event for webserver-dd94f59b7-6pgrf: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:39 +0000 UTC - event for webserver-dd94f59b7-2s9sc: {kubelet node2} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:42 +0000 UTC - event for webserver-dd94f59b7-5f2sj: {kubelet node1} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:46 +0000 UTC - event for webserver-5f9c9b577c-r2ztq: {kubelet node2} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:47 +0000 UTC - event for webserver-dd94f59b7-gf9ss: {kubelet node1} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:48 +0000 UTC - event for webserver-5f9c9b577c-q9rdk: {kubelet node1} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:49 +0000 UTC - event for webserver-5f9c9b577c-vl7j4: {kubelet node1} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:49 +0000 UTC - event for webserver-dd94f59b7-2lpbc: {kubelet node1} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.389: INFO: At 2021-05-12 19:38:49 +0000 UTC - event for webserver-dd94f59b7-ntg6v: {kubelet node1} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-976lc], unattached volumes=[default-token-976lc]: timed out waiting for the condition&#xA;May 12 19:42:08.391: INFO: POD                         NODE   PHASE    GRACE  CONDITIONS&#xA;May 12 19:42:08.391: INFO: webserver-5f9c9b577c-5bnpn  node1  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:37 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:37 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:37 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:37 +0000 UTC  }]&#xA;May 12 19:42:08.391: INFO: webserver-5f9c9b577c-zk25n  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:36 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:36 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:36 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:36 +0000 UTC  }]&#xA;May 12 19:42:08.391: INFO: webserver-6699947579-4rbmr  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC  }]&#xA;May 12 19:42:08.392: INFO: webserver-dd94f59b7-6pgrf   node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:37:06 +0000 UTC  }]&#xA;May 12 19:42:08.392: INFO: webserver-dd94f59b7-dll2m   node1  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:49 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:49 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:36:49 +0000 UTC  }]&#xA;May 12 19:42:08.392: INFO: &#xA;May 12 19:42:08.396: INFO: &#xA;Logging node info for node master1&#xA;May 12 19:42:08.398: INFO: Node Info: &amp;Node{ObjectMeta:{master1   /api/v1/nodes/master1 03e64d13-444c-41eb-b6bd-3745f05cd1cd 71348 0 2021-05-12 16:30:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;ce:94:4b:fc:cf:bc&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:30:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:30:45 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {kube-controller-manager Update v1 2021-05-12 16:31:10 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:27 +0000 UTC,LastTransitionTime:2021-05-12 16:35:27 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:07 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:07 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:07 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:42:07 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:201a06a0695b47fdbdd7df0fdd94f4dc,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:a7dddfd4-eda7-4a08-9750-d4a3c87c0cdd,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:60f4b5001bb5e7280fddf9143d3ed9bcde4e8016eef54522b5aea6bac9d9774b tas-controller:latest localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[nginx@sha256:a97eb9ecc708c8aa715ccfb5e9338f5456e4b65575daf304f108301f3b497314 nginx:1.19.2-alpine],SizeBytes:22052669,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:1935bf73835eb6d4446668a5484bb3724d97d926b26b41c7bf064aa3a5a8bc5f tas-extender:latest localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:42:08.398: INFO: &#xA;Logging kubelet events for node master1&#xA;May 12 19:42:08.400: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May 12 19:42:08.421: INFO: kube-scheduler-master1 started at 2021-05-12 16:32:00 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;May 12 19:42:08.422: INFO: kube-proxy-v5zpq started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:42:08.422: INFO: kube-flannel-rc2gh started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:42:08.422: INFO: &#x9;Container kube-flannel ready: true, restart count 3&#xA;May 12 19:42:08.422: INFO: coredns-7677f9bb54-4lkxs started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container coredns ready: true, restart count 2&#xA;May 12 19:42:08.422: INFO: docker-registry-docker-registry-56cbc7bc58-lfjnt started at 2021-05-12 16:36:16 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May 12 19:42:08.422: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;May 12 19:42:08.422: INFO: kube-apiserver-master1 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:42:08.422: INFO: kube-multus-ds-amd64-rcwwm started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:42:08.422: INFO: node-exporter-wd9j2 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.422: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:42:08.422: INFO: kube-controller-manager-master1 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.422: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;W0512 19:42:08.433302      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:42:08.465: INFO: &#xA;Latency metrics for node master1&#xA;May 12 19:42:08.465: INFO: &#xA;Logging node info for node master2&#xA;May 12 19:42:08.468: INFO: Node Info: &amp;Node{ObjectMeta:{master2   /api/v1/nodes/master2 f3d8b0a9-b7ca-4942-bfc7-69b182fddff8 71304 0 2021-05-12 16:31:23 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;32:9d:92:dd:24:1d&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:31:23 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:31:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-12 16:33:30 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:36:36 +0000 UTC,LastTransitionTime:2021-05-12 16:36:36 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:33:30 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:ef3726c355844802be21d744dd831d84,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:6feec048-83de-4ff4-ae36-12985bcda218,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:42:08.468: INFO: &#xA;Logging kubelet events for node master2&#xA;May 12 19:42:08.471: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May 12 19:42:08.479: INFO: kube-proxy-6ljsd started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:42:08.479: INFO: kube-flannel-pppch started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:42:08.479: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 12 19:42:08.479: INFO: kube-multus-ds-amd64-4swdk started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:42:08.479: INFO: coredns-7677f9bb54-b628b started at 2021-05-12 16:33:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May 12 19:42:08.479: INFO: node-exporter-dprww started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.479: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:42:08.479: INFO: kube-apiserver-master2 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:42:08.479: INFO: kube-controller-manager-master2 started at 2021-05-12 16:38:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:42:08.479: INFO: kube-scheduler-master2 started at 2021-05-12 16:32:00 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.479: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;W0512 19:42:08.503520      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:42:08.534: INFO: &#xA;Latency metrics for node master2&#xA;May 12 19:42:08.534: INFO: &#xA;Logging node info for node master3&#xA;May 12 19:42:08.537: INFO: Node Info: &amp;Node{ObjectMeta:{master3   /api/v1/nodes/master3 a0124635-fd55-4774-af26-32765b659025 71313 0 2021-05-12 16:31:34 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;8e:91:ce:de:d3:10&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:31:34 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:31:35 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-12 16:33:30 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {nfd-master Update v1 2021-05-12 16:39:03 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:42 +0000 UTC,LastTransitionTime:2021-05-12 16:35:42 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:42:06 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:8966dda31ac243c3b9142da7eb1a3315,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:29bf94f6-9eaa-4435-8e2a-f17b4721b8df,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:9d07c391aeb1a9d02eb4343c113ed01825227c70c32b3cae861711f90191b0fd quay.io/coreos/kube-rbac-proxy:v0.4.1],SizeBytes:41317870,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[quay.io/coreos/prometheus-operator@sha256:a54e806fb27d2fb0251da4f3b2a3bb5320759af63a54a755788304775f2384a7 quay.io/coreos/prometheus-operator:v0.40.0],SizeBytes:38238457,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:42:08.537: INFO: &#xA;Logging kubelet events for node master3&#xA;May 12 19:42:08.539: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May 12 19:42:08.548: INFO: node-feature-discovery-controller-5bf5c49849-gdz9z started at 2021-05-12 16:38:58 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;May 12 19:42:08.548: INFO: node-exporter-jzp84 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.548: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:42:08.548: INFO: kube-scheduler-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-scheduler ready: true, restart count 3&#xA;May 12 19:42:08.548: INFO: kube-proxy-d57dm started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 12 19:42:08.548: INFO: kube-flannel-hmbzn started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 12 19:42:08.548: INFO: dns-autoscaler-5b7b5c9b6f-8fzpk started at 2021-05-12 16:33:51 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;May 12 19:42:08.548: INFO: kube-apiserver-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:42:08.548: INFO: kube-controller-manager-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:42:08.548: INFO: kube-multus-ds-amd64-4hlf8 started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:42:08.548: INFO: prometheus-operator-5bb8cb9d8f-sqgl4 started at 2021-05-12 16:42:55 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.548: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.548: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;W0512 19:42:08.559619      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:42:08.583: INFO: &#xA;Latency metrics for node master3&#xA;May 12 19:42:08.583: INFO: &#xA;Logging node info for node node1&#xA;May 12 19:42:08.585: INFO: Node Info: &amp;Node{ObjectMeta:{node1   /api/v1/nodes/node1 2f7b1644-6865-412d-b423-194903813633 71349 0 2021-05-12 16:32:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;6a:b5:62:81:d9:8e&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:25 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-12 16:39:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-12 16:41:28 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:02:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:36:30 +0000 UTC,LastTransitionTime:2021-05-12 16:36:30 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:08 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:08 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:08 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:42:08 +0000 UTC,LastTransitionTime:2021-05-12 16:33:25 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:d9a9ef88c91340689744923e9951f78a,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:d2f244db-a37b-4dba-a8f9-6ea24c6cdb4e,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:96570ae3225d359c9667e2cbd57987388aa9d3be9d5a198ed9b677c7f9f4e450 localhost:30500/barometer-collectd:stable],SizeBytes:1464260582,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1002488025,},ContainerImage{Names:[opnfv/barometer-collectd@sha256:ed5c574f653e2a39e784ff322033a2319aafde7366c803a88f20f7a2a8bc1efb opnfv/barometer-collectd:stable],SizeBytes:825413035,},ContainerImage{Names:[localhost:30500/cmk@sha256:f178577724ef3b8118ca58277a27d594257de6c55a813b61e421df4008a7f73b cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:1636899c10870ab66c48d960a9df620f4f9e86a0c72fbacf36032d27404e7e6c golang:alpine3.12],SizeBytes:301156062,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[grafana/grafana@sha256:89304bc2335f4976618548d7b93d165ed67369d3a051d2f627fc4e0aa3d0aff1 grafana/grafana:7.1.0],SizeBytes:179601493,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:d4ba4dd1a9ebb90916d0bfed3c204adcb118ed24546bf8dd2e6b30fc0fd2009e quay.io/prometheus/prometheus:v2.20.0],SizeBytes:144886595,},ContainerImage{Names:[nginx@sha256:34ac17eb150f557f6ce94ea34e7c03f3899b717307cad633213a9488a431179f nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter-amd64@sha256:b63dc612e3cb73f79d2401a4516f794f9f0a83002600ca72e675e41baecff437 directxman12/k8s-prometheus-adapter-amd64:v0.6.0],SizeBytes:53267842,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:155361826d160d9f566aede3ea35b34f1e0c6422720285b1c22467c2b21a90aa nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392919,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[quay.io/coreos/prometheus-config-reloader@sha256:c679a143b24b7731ad1577a9865aa3805426cbf1b25e30807b951dff68466ffd quay.io/coreos/prometheus-config-reloader:v0.40.0],SizeBytes:10131705,},ContainerImage{Names:[jimmidyson/configmap-reload@sha256:d107c7a235c266273b1c3502a391fec374430e5625539403d0de797fa9c556a2 jimmidyson/configmap-reload:v0.3.0],SizeBytes:9700438,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:36553b10a4947067b9fbb7d532951066293a68eae893beba1d9235f7d11a20ad alpine:3.12],SizeBytes:5581415,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:42:08.586: INFO: &#xA;Logging kubelet events for node node1&#xA;May 12 19:42:08.588: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May 12 19:42:08.693: INFO: ss2-0 started at 2021-05-12 19:36:33 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container webserver ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: all-succeed-x6rzf started at 2021-05-12 19:36:55 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: concurrent-1620848520-6rh26 started at 2021-05-12 19:42:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: forbid-1620848520-52wz6 started at 2021-05-12 19:42:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: kubernetes-dashboard-86c6f9df5b-vkvbq started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 1&#xA;May 12 19:42:08.693: INFO: webserver-dd94f59b7-dll2m started at 2021-05-12 19:36:49 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: prometheus-k8s-0 started at 2021-05-12 16:43:20 +0000 UTC (0+5 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May 12 19:42:08.693: INFO: &#x9;Container prometheus-config-reloader ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container rules-configmap-reloader ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: collectd-5mpmz started at 2021-05-12 16:49:38 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: kube-proxy-r9vsx started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 12 19:42:08.693: INFO: node-exporter-ddxbd started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: webserver-5f9c9b577c-5bnpn started at 2021-05-12 19:36:37 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: kube-flannel-r7w6z started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Init container install-cni ready: true, restart count 1&#xA;May 12 19:42:08.693: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 12 19:42:08.693: INFO: kube-multus-ds-amd64-fhzwc started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:42:08.693: INFO: kubernetes-metrics-scraper-678c97765c-s4sgj started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May 12 19:42:08.693: INFO: cmk-v4qwz started at 2021-05-12 16:42:07 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-46jff started at 2021-05-12 16:39:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: cmk-init-discover-node1-2x2zk started at 2021-05-12 16:41:25 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 12 19:42:08.693: INFO: rs-92pbk started at 2021-05-12 19:42:01 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container donothing ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: rs-ckftr started at 2021-05-12 19:42:01 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container donothing ready: true, restart count 0&#xA;May 12 19:42:08.693: INFO: nginx-proxy-node1 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 12 19:42:08.693: INFO: node-feature-discovery-worker-qtn84 started at 2021-05-12 16:38:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.693: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;W0512 19:42:08.705324      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:42:08.951: INFO: &#xA;Latency metrics for node node1&#xA;May 12 19:42:08.951: INFO: &#xA;Logging node info for node node2&#xA;May 12 19:42:08.953: INFO: Node Info: &amp;Node{ObjectMeta:{node2   /api/v1/nodes/node2 16e4ef1f-0c51-44d5-a257-2eaa348c0d52 71141 0 2021-05-12 16:32:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;52:9a:ee:23:4c:ac&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:25 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-12 16:39:06 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-12 16:41:50 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {e2e.test Update v1 2021-05-12 19:08:41 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:example.com/fakecpu&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:08:51 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:example.com/fakecpu&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:14 +0000 UTC,LastTransitionTime:2021-05-12 16:35:14 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:01 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:01 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:42:01 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:42:01 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:eebaf4858fee4a739009f2f7f2717953,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:15bd8119-98f5-4e13-a2fa-897c2224dd6e,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:96570ae3225d359c9667e2cbd57987388aa9d3be9d5a198ed9b677c7f9f4e450 localhost:30500/barometer-collectd:stable],SizeBytes:1464260582,},ContainerImage{Names:[localhost:30500/cmk@sha256:f178577724ef3b8118ca58277a27d594257de6c55a813b61e421df4008a7f73b localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[nginx@sha256:34ac17eb150f557f6ce94ea34e7c03f3899b717307cad633213a9488a431179f nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/sample-apiserver@sha256:ff02aacd9766d597883fabafc7ad604c719a57611db1bcc1564c69a45b000a55 gcr.io/kubernetes-e2e-test-images/sample-apiserver:1.17],SizeBytes:60684726,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:155361826d160d9f566aede3ea35b34f1e0c6422720285b1c22467c2b21a90aa localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392919,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:60f4b5001bb5e7280fddf9143d3ed9bcde4e8016eef54522b5aea6bac9d9774b localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:1935bf73835eb6d4446668a5484bb3724d97d926b26b41c7bf064aa3a5a8bc5f localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[gcr.io/authenticated-image-pulling/alpine@sha256:7ff177862cb50c602bfe81f805969412e619c054a2bbead977d0c276988aa4a0 gcr.io/authenticated-image-pulling/alpine:3.7],SizeBytes:4206620,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:42:08.954: INFO: &#xA;Logging kubelet events for node node2&#xA;May 12 19:42:08.957: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May 12 19:42:08.971: INFO: kube-multus-ds-amd64-k28rf started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:42:08.971: INFO: node-feature-discovery-worker-zjvzk started at 2021-05-12 16:38:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-j6r4r started at 2021-05-12 16:39:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: node-exporter-h5rv7 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: test-orphan-deployment-dd94f59b7-phg85 started at 2021-05-12 19:41:37 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: tas-telemetry-aware-scheduling-575ccbc9d4-8xc75 started at 2021-05-12 16:46:05 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container tas-controller ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: pod-0 started at 2021-05-12 19:42:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container donothing ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: nginx-proxy-node2 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 12 19:42:08.971: INFO: kube-flannel-rqtcs started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:42:08.971: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 12 19:42:08.971: INFO: cmk-webhook-6c9d5f8578-fgcvr started at 2021-05-12 16:42:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: rs-457dt started at 2021-05-12 19:42:01 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container donothing ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: cmk-init-discover-node2-qrd9v started at 2021-05-12 16:41:44 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: pod-1 started at 2021-05-12 19:42:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container donothing ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: webserver-dd94f59b7-6pgrf started at 2021-05-12 19:37:06 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: all-succeed-rbrzg started at 2021-05-12 19:36:55 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: kube-proxy-grtqc started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:42:08.971: INFO: cmk-gbw5d started at 2021-05-12 16:42:08 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: backofflimit-vrd8v started at 2021-05-12 19:36:52 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: collectd-tng6x started at 2021-05-12 16:49:38 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 12 19:42:08.971: INFO: all-pods-removed-r4vfz started at 2021-05-12 19:37:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: webserver-6699947579-4rbmr started at 2021-05-12 19:37:06 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: webserver-5f9c9b577c-zk25n started at 2021-05-12 19:36:36 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 12 19:42:08.971: INFO: all-pods-removed-mzshw started at 2021-05-12 19:37:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:42:08.971: INFO: &#x9;Container c ready: false, restart count 0&#xA;W0512 19:42:08.984573      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:42:09.040: INFO: &#xA;Latency metrics for node node2&#xA;May 12 19:42:09.041: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;deployment-4463&#34; for this suite.&#xA;</system-out>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-bindmounted] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PV Protection Verify &#34;immediate&#34; deletion of a PV that is not bound to a PVC" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeFeature:FSGroup]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Logging soak [Performance] [Slow] [Disruptive] should survive logging 1KB every 1s seconds, for a duration of 2m0s" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Network Partition [Disruptive] [Slow] [k8s.io] Pods should return to running and ready state after network partition is healed All pods on the unreachable node should be marked as NotReady upon the node turn NotReady AND all pods should be mark back to Ready when the node get back to Ready before pod eviction timeout" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should provide podname as non-root with fsgroup [LinuxOnly] [NodeFeature:FSGroup]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should provide basic identity" classname="Kubernetes e2e suite" time="0.046927828">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Job should run a job to completion when tasks sometimes fail and are not locally restarted" classname="Kubernetes e2e suite" time="900.391928191">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:117&#xA;May 12 19:57:09.266: failed to ensure job completion in namespace: job-503&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0003021f0&gt;: {&#xA;        s: &#34;timed out waiting for the condition&#34;,&#xA;    }&#xA;    timed out waiting for the condition&#xA;occurred&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:136</failure>
          <system-out>[BeforeEach] [sig-apps] Job&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174&#xA;STEP: Creating a kubernetes client&#xA;May 12 19:42:09.231: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename job&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[It] should run a job to completion when tasks sometimes fail and are not locally restarted&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:117&#xA;STEP: Looking for a node to schedule job pod&#xA;STEP: Creating a job&#xA;STEP: Ensuring job reaches completions&#xA;May 12 19:57:09.266: FAIL: failed to ensure job completion in namespace: job-503&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0003021f0&gt;: {&#xA;        s: &#34;timed out waiting for the condition&#34;,&#xA;    }&#xA;    timed out waiting for the condition&#xA;occurred&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/apps.glob..func6.4()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:136 +0x3e9&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc000601c80)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc000601c80)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc000601c80, 0x4de37a0)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;[AfterEach] [sig-apps] Job&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175&#xA;STEP: Collecting events from namespace &#34;job-503&#34;.&#xA;STEP: Found 19 events.&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:09 +0000 UTC - event for fail-once-non-local: {job-controller } SuccessfulCreate: Created pod: fail-once-non-local-2vxhb&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:09 +0000 UTC - event for fail-once-non-local: {job-controller } SuccessfulCreate: Created pod: fail-once-non-local-dl6pl&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:13 +0000 UTC - event for fail-once-non-local-2vxhb: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:13 +0000 UTC - event for fail-once-non-local-2vxhb: {multus } AddedInterface: Add eth0 [10.244.3.22/24]&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:13 +0000 UTC - event for fail-once-non-local-dl6pl: {multus } AddedInterface: Add eth0 [10.244.3.19/24]&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:13 +0000 UTC - event for fail-once-non-local-dl6pl: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:14 +0000 UTC - event for fail-once-non-local-2vxhb: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:14 +0000 UTC - event for fail-once-non-local-2vxhb: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:14 +0000 UTC - event for fail-once-non-local-2vxhb: {kubelet node1} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:17 +0000 UTC - event for fail-once-non-local-2vxhb: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:17 +0000 UTC - event for fail-once-non-local-2vxhb: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:17 +0000 UTC - event for fail-once-non-local-2vxhb: {multus } AddedInterface: Add eth0 [10.244.3.23/24]&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:17 +0000 UTC - event for fail-once-non-local-dl6pl: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:17 +0000 UTC - event for fail-once-non-local-dl6pl: {kubelet node1} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:17 +0000 UTC - event for fail-once-non-local-dl6pl: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:20 +0000 UTC - event for fail-once-non-local-dl6pl: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:20 +0000 UTC - event for fail-once-non-local-dl6pl: {multus } AddedInterface: Add eth0 [10.244.3.25/24]&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:20 +0000 UTC - event for fail-once-non-local-dl6pl: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 12 19:57:09.272: INFO: At 2021-05-12 19:42:22 +0000 UTC - event for fail-once-non-local-2vxhb: {multus } AddedInterface: Add eth0 [10.244.3.26/24]&#xA;May 12 19:57:09.274: INFO: POD                        NODE   PHASE    GRACE  CONDITIONS&#xA;May 12 19:57:09.274: INFO: fail-once-non-local-2vxhb  node1  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC  }]&#xA;May 12 19:57:09.274: INFO: fail-once-non-local-dl6pl  node1  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:09 +0000 UTC  }]&#xA;May 12 19:57:09.274: INFO: &#xA;May 12 19:57:09.278: INFO: &#xA;Logging node info for node master1&#xA;May 12 19:57:09.280: INFO: Node Info: &amp;Node{ObjectMeta:{master1   /api/v1/nodes/master1 03e64d13-444c-41eb-b6bd-3745f05cd1cd 75556 0 2021-05-12 16:30:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;ce:94:4b:fc:cf:bc&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:30:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:30:45 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {kube-controller-manager Update v1 2021-05-12 16:31:10 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:27 +0000 UTC,LastTransitionTime:2021-05-12 16:35:27 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:01 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:01 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:01 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:57:01 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:201a06a0695b47fdbdd7df0fdd94f4dc,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:a7dddfd4-eda7-4a08-9750-d4a3c87c0cdd,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:60f4b5001bb5e7280fddf9143d3ed9bcde4e8016eef54522b5aea6bac9d9774b tas-controller:latest localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[nginx@sha256:a97eb9ecc708c8aa715ccfb5e9338f5456e4b65575daf304f108301f3b497314 nginx:1.19.2-alpine],SizeBytes:22052669,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:1935bf73835eb6d4446668a5484bb3724d97d926b26b41c7bf064aa3a5a8bc5f tas-extender:latest localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:57:09.281: INFO: &#xA;Logging kubelet events for node master1&#xA;May 12 19:57:09.283: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May 12 19:57:09.299: INFO: kube-apiserver-master1 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:57:09.299: INFO: kube-scheduler-master1 started at 2021-05-12 16:32:00 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;May 12 19:57:09.299: INFO: kube-proxy-v5zpq started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:57:09.299: INFO: kube-flannel-rc2gh started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:57:09.299: INFO: &#x9;Container kube-flannel ready: true, restart count 3&#xA;May 12 19:57:09.299: INFO: coredns-7677f9bb54-4lkxs started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container coredns ready: true, restart count 2&#xA;May 12 19:57:09.299: INFO: docker-registry-docker-registry-56cbc7bc58-lfjnt started at 2021-05-12 16:36:16 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May 12 19:57:09.299: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;May 12 19:57:09.299: INFO: kube-controller-manager-master1 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:57:09.299: INFO: kube-multus-ds-amd64-rcwwm started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:57:09.299: INFO: node-exporter-wd9j2 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.299: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.299: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;W0512 19:57:09.313467      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:57:09.339: INFO: &#xA;Latency metrics for node master1&#xA;May 12 19:57:09.339: INFO: &#xA;Logging node info for node master2&#xA;May 12 19:57:09.341: INFO: Node Info: &amp;Node{ObjectMeta:{master2   /api/v1/nodes/master2 f3d8b0a9-b7ca-4942-bfc7-69b182fddff8 75548 0 2021-05-12 16:31:23 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;32:9d:92:dd:24:1d&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:31:23 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:31:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-12 16:33:30 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:36:36 +0000 UTC,LastTransitionTime:2021-05-12 16:36:36 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:56:59 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:56:59 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:56:59 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:56:59 +0000 UTC,LastTransitionTime:2021-05-12 16:33:30 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:ef3726c355844802be21d744dd831d84,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:6feec048-83de-4ff4-ae36-12985bcda218,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:57:09.342: INFO: &#xA;Logging kubelet events for node master2&#xA;May 12 19:57:09.346: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May 12 19:57:09.361: INFO: node-exporter-dprww started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.361: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:57:09.361: INFO: kube-apiserver-master2 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:57:09.361: INFO: kube-controller-manager-master2 started at 2021-05-12 16:38:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:57:09.361: INFO: kube-scheduler-master2 started at 2021-05-12 16:32:00 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;May 12 19:57:09.361: INFO: kube-proxy-6ljsd started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:57:09.361: INFO: kube-flannel-pppch started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:57:09.361: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 12 19:57:09.361: INFO: kube-multus-ds-amd64-4swdk started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:57:09.361: INFO: coredns-7677f9bb54-b628b started at 2021-05-12 16:33:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.361: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;W0512 19:57:09.375139      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:57:09.399: INFO: &#xA;Latency metrics for node master2&#xA;May 12 19:57:09.399: INFO: &#xA;Logging node info for node master3&#xA;May 12 19:57:09.401: INFO: Node Info: &amp;Node{ObjectMeta:{master3   /api/v1/nodes/master3 a0124635-fd55-4774-af26-32765b659025 75554 0 2021-05-12 16:31:34 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;8e:91:ce:de:d3:10&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:31:34 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:31:35 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-12 16:33:30 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {nfd-master Update v1 2021-05-12 16:39:03 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:42 +0000 UTC,LastTransitionTime:2021-05-12 16:35:42 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:00 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:00 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:00 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:57:00 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:8966dda31ac243c3b9142da7eb1a3315,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:29bf94f6-9eaa-4435-8e2a-f17b4721b8df,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:9d07c391aeb1a9d02eb4343c113ed01825227c70c32b3cae861711f90191b0fd quay.io/coreos/kube-rbac-proxy:v0.4.1],SizeBytes:41317870,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[quay.io/coreos/prometheus-operator@sha256:a54e806fb27d2fb0251da4f3b2a3bb5320759af63a54a755788304775f2384a7 quay.io/coreos/prometheus-operator:v0.40.0],SizeBytes:38238457,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:57:09.401: INFO: &#xA;Logging kubelet events for node master3&#xA;May 12 19:57:09.404: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May 12 19:57:09.418: INFO: kube-apiserver-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:57:09.418: INFO: kube-controller-manager-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:57:09.418: INFO: kube-multus-ds-amd64-4hlf8 started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:57:09.418: INFO: prometheus-operator-5bb8cb9d8f-sqgl4 started at 2021-05-12 16:42:55 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.418: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;May 12 19:57:09.418: INFO: node-exporter-jzp84 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.418: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:57:09.418: INFO: kube-scheduler-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-scheduler ready: true, restart count 3&#xA;May 12 19:57:09.418: INFO: kube-proxy-d57dm started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 12 19:57:09.418: INFO: kube-flannel-hmbzn started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:57:09.418: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 12 19:57:09.418: INFO: dns-autoscaler-5b7b5c9b6f-8fzpk started at 2021-05-12 16:33:51 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;May 12 19:57:09.418: INFO: node-feature-discovery-controller-5bf5c49849-gdz9z started at 2021-05-12 16:38:58 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.418: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;W0512 19:57:09.430394      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:57:09.461: INFO: &#xA;Latency metrics for node master3&#xA;May 12 19:57:09.461: INFO: &#xA;Logging node info for node node1&#xA;May 12 19:57:09.463: INFO: Node Info: &amp;Node{ObjectMeta:{node1   /api/v1/nodes/node1 2f7b1644-6865-412d-b423-194903813633 75565 0 2021-05-12 16:32:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;6a:b5:62:81:d9:8e&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:25 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-12 16:39:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-12 16:41:28 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:02:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:36:30 +0000 UTC,LastTransitionTime:2021-05-12 16:36:30 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:04 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:04 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:04 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:57:04 +0000 UTC,LastTransitionTime:2021-05-12 16:33:25 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:d9a9ef88c91340689744923e9951f78a,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:d2f244db-a37b-4dba-a8f9-6ea24c6cdb4e,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:96570ae3225d359c9667e2cbd57987388aa9d3be9d5a198ed9b677c7f9f4e450 localhost:30500/barometer-collectd:stable],SizeBytes:1464260582,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1002488025,},ContainerImage{Names:[opnfv/barometer-collectd@sha256:ed5c574f653e2a39e784ff322033a2319aafde7366c803a88f20f7a2a8bc1efb opnfv/barometer-collectd:stable],SizeBytes:825413035,},ContainerImage{Names:[localhost:30500/cmk@sha256:f178577724ef3b8118ca58277a27d594257de6c55a813b61e421df4008a7f73b cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:1636899c10870ab66c48d960a9df620f4f9e86a0c72fbacf36032d27404e7e6c golang:alpine3.12],SizeBytes:301156062,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[grafana/grafana@sha256:89304bc2335f4976618548d7b93d165ed67369d3a051d2f627fc4e0aa3d0aff1 grafana/grafana:7.1.0],SizeBytes:179601493,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:d4ba4dd1a9ebb90916d0bfed3c204adcb118ed24546bf8dd2e6b30fc0fd2009e quay.io/prometheus/prometheus:v2.20.0],SizeBytes:144886595,},ContainerImage{Names:[nginx@sha256:34ac17eb150f557f6ce94ea34e7c03f3899b717307cad633213a9488a431179f nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter-amd64@sha256:b63dc612e3cb73f79d2401a4516f794f9f0a83002600ca72e675e41baecff437 directxman12/k8s-prometheus-adapter-amd64:v0.6.0],SizeBytes:53267842,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:155361826d160d9f566aede3ea35b34f1e0c6422720285b1c22467c2b21a90aa nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392919,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[quay.io/coreos/prometheus-config-reloader@sha256:c679a143b24b7731ad1577a9865aa3805426cbf1b25e30807b951dff68466ffd quay.io/coreos/prometheus-config-reloader:v0.40.0],SizeBytes:10131705,},ContainerImage{Names:[jimmidyson/configmap-reload@sha256:d107c7a235c266273b1c3502a391fec374430e5625539403d0de797fa9c556a2 jimmidyson/configmap-reload:v0.3.0],SizeBytes:9700438,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:36553b10a4947067b9fbb7d532951066293a68eae893beba1d9235f7d11a20ad alpine:3.12],SizeBytes:5581415,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:57:09.464: INFO: &#xA;Logging kubelet events for node node1&#xA;May 12 19:57:09.467: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May 12 19:57:09.487: INFO: kubernetes-metrics-scraper-678c97765c-s4sgj started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.487: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May 12 19:57:09.487: INFO: cmk-v4qwz started at 2021-05-12 16:42:07 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.487: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 12 19:57:09.487: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 12 19:57:09.487: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-46jff started at 2021-05-12 16:39:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.487: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 12 19:57:09.487: INFO: cmk-init-discover-node1-2x2zk started at 2021-05-12 16:41:25 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:57:09.487: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 12 19:57:09.488: INFO: fail-once-non-local-2vxhb started at 2021-05-12 19:42:09 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:57:09.488: INFO: nginx-proxy-node1 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 12 19:57:09.488: INFO: node-feature-discovery-worker-qtn84 started at 2021-05-12 16:38:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: kubernetes-dashboard-86c6f9df5b-vkvbq started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 1&#xA;May 12 19:57:09.488: INFO: prometheus-k8s-0 started at 2021-05-12 16:43:20 +0000 UTC (0+5 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May 12 19:57:09.488: INFO: &#x9;Container prometheus-config-reloader ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container rules-configmap-reloader ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: collectd-5mpmz started at 2021-05-12 16:49:38 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: kube-proxy-r9vsx started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 12 19:57:09.488: INFO: node-exporter-ddxbd started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:57:09.488: INFO: fail-once-non-local-dl6pl started at 2021-05-12 19:42:09 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:57:09.488: INFO: kube-flannel-r7w6z started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Init container install-cni ready: true, restart count 1&#xA;May 12 19:57:09.488: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 12 19:57:09.488: INFO: kube-multus-ds-amd64-fhzwc started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.488: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;W0512 19:57:09.500400      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:57:09.545: INFO: &#xA;Latency metrics for node node1&#xA;May 12 19:57:09.545: INFO: &#xA;Logging node info for node node2&#xA;May 12 19:57:09.548: INFO: Node Info: &amp;Node{ObjectMeta:{node2   /api/v1/nodes/node2 16e4ef1f-0c51-44d5-a257-2eaa348c0d52 75572 0 2021-05-12 16:32:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;52:9a:ee:23:4c:ac&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:25 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-12 16:39:06 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-12 16:41:50 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {e2e.test Update v1 2021-05-12 19:08:41 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:example.com/fakecpu&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:08:51 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:example.com/fakecpu&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:14 +0000 UTC,LastTransitionTime:2021-05-12 16:35:14 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:06 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:06 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:57:06 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:57:06 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:eebaf4858fee4a739009f2f7f2717953,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:15bd8119-98f5-4e13-a2fa-897c2224dd6e,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:96570ae3225d359c9667e2cbd57987388aa9d3be9d5a198ed9b677c7f9f4e450 localhost:30500/barometer-collectd:stable],SizeBytes:1464260582,},ContainerImage{Names:[localhost:30500/cmk@sha256:f178577724ef3b8118ca58277a27d594257de6c55a813b61e421df4008a7f73b localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[nginx@sha256:34ac17eb150f557f6ce94ea34e7c03f3899b717307cad633213a9488a431179f nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/sample-apiserver@sha256:ff02aacd9766d597883fabafc7ad604c719a57611db1bcc1564c69a45b000a55 gcr.io/kubernetes-e2e-test-images/sample-apiserver:1.17],SizeBytes:60684726,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:155361826d160d9f566aede3ea35b34f1e0c6422720285b1c22467c2b21a90aa localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392919,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:60f4b5001bb5e7280fddf9143d3ed9bcde4e8016eef54522b5aea6bac9d9774b localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:1935bf73835eb6d4446668a5484bb3724d97d926b26b41c7bf064aa3a5a8bc5f localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[gcr.io/authenticated-image-pulling/alpine@sha256:7ff177862cb50c602bfe81f805969412e619c054a2bbead977d0c276988aa4a0 gcr.io/authenticated-image-pulling/alpine:3.7],SizeBytes:4206620,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:57:09.549: INFO: &#xA;Logging kubelet events for node node2&#xA;May 12 19:57:09.551: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May 12 19:57:09.569: INFO: collectd-tng6x started at 2021-05-12 16:49:38 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-j6r4r started at 2021-05-12 16:39:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: node-exporter-h5rv7 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: kube-multus-ds-amd64-k28rf started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:57:09.569: INFO: node-feature-discovery-worker-zjvzk started at 2021-05-12 16:38:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: tas-telemetry-aware-scheduling-575ccbc9d4-8xc75 started at 2021-05-12 16:46:05 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container tas-controller ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: cmk-webhook-6c9d5f8578-fgcvr started at 2021-05-12 16:42:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: nginx-proxy-node2 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 12 19:57:09.569: INFO: kube-flannel-rqtcs started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:57:09.569: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 12 19:57:09.569: INFO: cmk-init-discover-node2-qrd9v started at 2021-05-12 16:41:44 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 12 19:57:09.569: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 12 19:57:09.569: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 12 19:57:09.569: INFO: kube-proxy-grtqc started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:57:09.569: INFO: cmk-gbw5d started at 2021-05-12 16:42:08 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:57:09.569: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 12 19:57:09.569: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;W0512 19:57:09.583406      37 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:57:09.614: INFO: &#xA;Latency metrics for node node2&#xA;May 12 19:57:09.614: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;job-503&#34; for this suite.&#xA;</system-out>
      </testcase>
  </testsuite>