<?xml version="1.0" encoding="UTF-8"?>
  <testsuite name="Kubernetes e2e suite" tests="2" failures="1" errors="0" time="356.65">
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI attach test using mock driver should preserve attachment policy when no CSIDriver present" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Character Device [Slow] Should fail on mounting non-existent character device &#39;does-not-exist-char-dev&#39; when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce policy based on PodSelector with MatchExpressions[Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should work after restarting kube-proxy [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning DynamicProvisioner External should let an external dynamic provisioner create and delete persistent volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Network Partition [Disruptive] [Slow] [k8s.io] [StatefulSet] should not reschedule stateful pods if there is a network partition [Slow] [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:Ingress] should support multiple TLS certs" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType File [Slow] Should fail on mounting file &#39;afile&#39; when HostPathType is HostPathBlockDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link-bindmounted] One pod requesting one prebound PVC should be able to mount volume and write from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] Should not scale GPU pool up if pod does not require GPUs [GpuType:] [Feature:ClusterSizeAutoscalingGpu]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link-bindmounted] Set fsGroup for local volume should set fsGroup for one pod [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Provisioning On Clustered Datastore [Feature:vsphere] verify static provisioning on clustered datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: block] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] ReplicaSet should serve a basic image on each replica with a public image  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [sig-autoscaling] [Serial] [Slow] Deployment Should scale from 5 pods to 3 pods and from 3 to 1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cluster-lifecycle] [Feature:BootstrapTokens] should delete the signed bootstrap tokens from clusterInfo ConfigMap when bootstrap token is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes:vsphere [Feature:vsphere] should test that deleting the PV before the pod does not cause pod deletion to fail on vsphere volume detach" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-cloud-provider-gcp] gpu Upgrade [Feature:GPUUpgrade] cluster upgrade should be able to run gpu pod after upgrade [Feature:GPUClusterUpgrade]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes Default StorageClass pods that use multiple volumes should be reschedulable [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI Volume expansion should not expand volume if resizingOnDriver=off, resizingOnSC=on" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV with storage policy specified in storage class in waitForFirstConsumer binding mode with multiple allowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] files with FSGroup ownership should support (root,0644,tmpfs)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] Should scale down GPU pool from 1 [GpuType:] [Feature:ClusterSizeAutoscalingGpu]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Stackdriver Metadata Agent [Feature:StackdriverMetadataAgent]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a ClusterIP service" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] PodSecurityPolicy [Feature:PodSecurityPolicy] should allow pods under the privileged policy.PodSecurityPolicy" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory &#39;adir&#39; when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce policy to allow traffic from pods within server namespace based on PodSelector [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Network Partition [Disruptive] [Slow] [k8s.io] Pods should return to running and ready state after network partition is healed All pods on the unreachable node should be marked as NotReady upon the node turn NotReady AND all pods should be mark back to Ready when the node get back to Ready before pod eviction timeout" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should have session affinity work for LoadBalancer service with ESIPP on [Slow] [DisabledForLargeClusters] [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Firewall rule [Slow] [Serial] should create valid firewall rules for LoadBalancer type service" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Docker Containers should be able to override the image&#39;s default arguments (docker cmd) [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Socket [Slow] Should fail on mounting socket &#39;asocket&#39; when HostPathType is HostPathBlockDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Storage Policy Based Volume Provisioning [Feature:vsphere] verify an if a SPBM policy and VSAN capabilities cannot be honored for dynamically provisioned pvc using storageclass" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl apply should apply a new configuration to an existing RC" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Custom Metrics - Stackdriver Adapter for external metrics [Feature:StackdriverExternalMetrics]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DisruptionController evictions: enough pods, replicaSet, percentage =&gt; should allow an eviction" classname="Kubernetes e2e suite" time="20.065702197"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Placement [Feature:vsphere] should create and delete pod with multiple volumes from different datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should provide container&#39;s memory limit [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Secrets should be immutable if `immutable` field is set" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce egress policy allowing traffic to a server in a different namespace based on PodSelector and NamespaceSelector [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DisruptionController evictions: maxUnavailable deny evictions, integer =&gt; should not allow an eviction [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType File [Slow] Should be able to mount file &#39;afile&#39; successfully when HostPathType is HostPathFile" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Custom Metrics - Stackdriver Adapter for new resource model [Feature:StackdriverCustomMetrics]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Security Context should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning DynamicProvisioner [Slow] deletion should be idempotent" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable[Feature:VolumeSnapshotDataSource] volume snapshot controller  should check snapshot fields, check restore correctly works after modifying source data, check deletion" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device &#39;ablkdev&#39; when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPath should support r/w [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPath should support subPath [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] NFSPersistentVolumes[Disruptive][Flaky] when kubelet restarts Should test that a volume mounted to a pod that is force deleted while the kubelet is down unmounts when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] GenericPersistentVolume[Disruptive] When kubelet restarts Should test that a volume mounted to a pod that is deleted while the kubelet is down unmounts when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Namespaces [Serial] should delete fast enough (90 percent of 100 namespaces in 150 seconds)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: gce-localssd-scsi-fs] [Serial] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Character Device [Slow] Should fail on mounting character device &#39;achardev&#39; when HostPathType is HostPathSocket" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] Windows volume mounts  check volume mount permissions container should have readOnly permissions on hostMapPath" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 should support forwarding over websockets" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable[Feature:VolumeSnapshotDataSource] volume snapshot controller  should check snapshot fields, check restore correctly works after modifying source data, check deletion" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap should be immutable if `immutable` field is set" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify PVC creation fails if the availability zone specified in the storage class have no shared datastores under it." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) [Feature:GenericEphemeralVolume] (immediate-binding)] ephemeral should support multiple inline ephemeral volumes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable[Feature:VolumeSnapshotDataSource] volume snapshot controller  should check snapshot fields, check restore correctly works after modifying source data, check deletion" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by switching off the network interface and ensure they function upon switch on" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected downwardAPI should provide container&#39;s memory limit [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Deployment iterative rollouts should eventually progress" classname="Kubernetes e2e suite" time="335.307288622">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:121&#xA;May 14 23:09:23.775: Unexpected error:&#xA;    &lt;*errors.errorString | 0xc001255dc0&gt;: {&#xA;        s: &#34;error waiting for deployment \&#34;webserver\&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:\&#34;Available\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;MinimumReplicasUnavailable\&#34;, Message:\&#34;Deployment does not have minimum availability.\&#34;}, v1.DeploymentCondition{Type:\&#34;Progressing\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;ProgressDeadlineExceeded\&#34;, Message:\&#34;ReplicaSet \\\&#34;webserver-7d5989497\\\&#34; has timed out progressing.\&#34;}}, CollisionCount:(*int32)(nil)}&#34;,&#xA;    }&#xA;    error waiting for deployment &#34;webserver&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;occurred&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:648</failure>
          <system-out>[BeforeEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174&#xA;STEP: Creating a kubernetes client&#xA;May 14 23:03:48.834: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename deployment&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[BeforeEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:78&#xA;[It] iterative rollouts should eventually progress&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:121&#xA;May 14 23:03:48.855: INFO: Creating deployment &#34;webserver&#34;&#xA;May 14 23:03:48.861: INFO: 00: triggering a new rollout for deployment &#34;webserver&#34;&#xA;May 14 23:03:48.863: INFO: 00: scaling up&#xA;May 14 23:03:48.867: INFO: Updating deployment webserver&#xA;May 14 23:03:48.867: INFO: 01: resuming deployment &#34;webserver&#34;&#xA;May 14 23:03:48.872: INFO: 01: scaling up&#xA;May 14 23:03:48.875: INFO: Updating deployment webserver&#xA;May 14 23:03:49.192: INFO: 02: resuming deployment &#34;webserver&#34;&#xA;May 14 23:03:49.199: INFO: Updating deployment webserver&#xA;May 14 23:03:50.406: INFO: 03: arbitrarily deleting one or more deployment pods for deployment &#34;webserver&#34;&#xA;May 14 23:03:50.410: INFO: 03: deleting deployment pod &#34;webserver-7746d44bfb-d6p9c&#34;&#xA;May 14 23:03:50.417: INFO: 03: deleting deployment pod &#34;webserver-7746d44bfb-h5wnd&#34;&#xA;May 14 23:03:50.425: INFO: 03: deleting deployment pod &#34;webserver-dd94f59b7-g8jb5&#34;&#xA;May 14 23:03:50.431: INFO: 03: deleting deployment pod &#34;webserver-dd94f59b7-lxk7s&#34;&#xA;May 14 23:03:51.750: INFO: 04: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 14 23:03:51.757: INFO: Updating deployment webserver&#xA;May 14 23:03:53.568: INFO: 05: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 14 23:03:53.575: INFO: Updating deployment webserver&#xA;May 14 23:03:53.575: INFO: 06: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 14 23:03:55.593: INFO: Updating deployment webserver&#xA;May 14 23:03:55.593: INFO: 07: resuming deployment &#34;webserver&#34;&#xA;May 14 23:03:55.596: INFO: 07: scaling down&#xA;May 14 23:03:55.599: INFO: Updating deployment webserver&#xA;May 14 23:03:55.599: INFO: 08: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 14 23:03:57.612: INFO: Updating deployment webserver&#xA;May 14 23:04:00.663: INFO: 09: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 14 23:04:00.670: INFO: Updating deployment webserver&#xA;May 14 23:04:00.671: INFO: 10: resuming deployment &#34;webserver&#34;&#xA;May 14 23:04:00.673: INFO: 10: scaling down&#xA;May 14 23:04:02.682: INFO: Updating deployment webserver&#xA;May 14 23:04:05.309: INFO: 11: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 14 23:04:05.316: INFO: Updating deployment webserver&#xA;May 14 23:04:05.316: INFO: 12: triggering a new rollout for deployment &#34;webserver&#34;&#xA;May 14 23:04:05.318: INFO: 12: scaling up&#xA;May 14 23:04:07.331: INFO: Updating deployment webserver&#xA;May 14 23:04:07.331: INFO: 13: scaling deployment &#34;webserver&#34;&#xA;May 14 23:04:07.334: INFO: 13: scaling up&#xA;May 14 23:04:09.341: INFO: 13: scaling down&#xA;May 14 23:04:09.344: INFO: Updating deployment webserver&#xA;May 14 23:04:12.713: INFO: 14: scaling deployment &#34;webserver&#34;&#xA;May 14 23:04:12.719: INFO: Updating deployment webserver&#xA;May 14 23:04:12.719: INFO: 15: resuming deployment &#34;webserver&#34;&#xA;May 14 23:04:12.722: INFO: 15: scaling down&#xA;May 14 23:04:12.725: INFO: Updating deployment webserver&#xA;May 14 23:04:13.035: INFO: 16: resuming deployment &#34;webserver&#34;&#xA;May 14 23:04:13.040: INFO: Updating deployment webserver&#xA;May 14 23:04:17.076: INFO: 17: triggering a new rollout for deployment &#34;webserver&#34;&#xA;May 14 23:04:17.078: INFO: 17: scaling down&#xA;May 14 23:04:17.082: INFO: Updating deployment webserver&#xA;May 14 23:04:19.740: INFO: 18: rolling back a rollout for deployment &#34;webserver&#34;&#xA;May 14 23:04:19.747: INFO: Updating deployment webserver&#xA;May 14 23:04:19.747: INFO: 19: resuming deployment &#34;webserver&#34;&#xA;May 14 23:04:21.753: INFO: 19: scaling down&#xA;May 14 23:04:21.757: INFO: Updating deployment webserver&#xA;May 14 23:04:21.760: INFO: Waiting for deployment &#34;webserver&#34; to be observed by the controller&#xA;May 14 23:04:23.765: INFO: Waiting for deployment &#34;webserver&#34; status&#xA;May 14 23:04:23.767: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:25.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:27.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:29.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:31.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:33.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:35.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:37.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:39.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:41.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:43.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:45.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:47.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;True&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630257, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ReplicaSetUpdated&#34;, Message:&#34;ReplicaSet \&#34;webserver-5dc9cd4855\&#34; is progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:49.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:51.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:53.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:55.775: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:57.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:04:59.773: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:01.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:03.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:05.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:07.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:09.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:11.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:13.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:15.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:17.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:19.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:21.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:23.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:25.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:27.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:29.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:31.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:33.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:35.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:37.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:39.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:41.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:43.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:45.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:47.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:49.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:51.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:53.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:55.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:57.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:05:59.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:01.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:03.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:05.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:07.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:09.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:11.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:13.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:15.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:17.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:19.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:21.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:23.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:25.773: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:27.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:29.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:31.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:33.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:35.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:37.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:39.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:41.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:43.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:45.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:47.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:49.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:51.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:53.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:55.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:57.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:06:59.773: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:01.774: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:03.773: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:05.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:07.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:09.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:11.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:13.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:15.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:17.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:19.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:21.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:23.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:25.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:27.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:29.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:31.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:33.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:35.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:37.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:39.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:41.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:43.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:45.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:47.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:49.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:51.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:53.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:55.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:57.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:07:59.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:01.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:03.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:05.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:07.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:09.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:11.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:13.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:15.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:17.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:19.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:21.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:23.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:25.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:27.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:29.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:31.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:33.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:35.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:37.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:39.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:41.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:43.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:45.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:47.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:49.773: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:51.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:53.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:55.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:57.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:08:59.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:01.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:03.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:05.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:07.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:09.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:11.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:13.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:15.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:17.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:19.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:21.773: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:23.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:23.775: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;May 14 23:09:23.775: FAIL: Unexpected error:&#xA;    &lt;*errors.errorString | 0xc001255dc0&gt;: {&#xA;        s: &#34;error waiting for deployment \&#34;webserver\&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:\&#34;Available\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;MinimumReplicasUnavailable\&#34;, Message:\&#34;Deployment does not have minimum availability.\&#34;}, v1.DeploymentCondition{Type:\&#34;Progressing\&#34;, Status:\&#34;False\&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:\&#34;ProgressDeadlineExceeded\&#34;, Message:\&#34;ReplicaSet \\\&#34;webserver-7d5989497\\\&#34; has timed out progressing.\&#34;}}, CollisionCount:(*int32)(nil)}&#34;,&#xA;    }&#xA;    error waiting for deployment &#34;webserver&#34; status to match expectation: deployment status: v1.DeploymentStatus{ObservedGeneration:23, Replicas:4, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:&#34;Available&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630228, loc:(*time.Location)(0x770c940)}}, Reason:&#34;MinimumReplicasUnavailable&#34;, Message:&#34;Deployment does not have minimum availability.&#34;}, v1.DeploymentCondition{Type:&#34;Progressing&#34;, Status:&#34;False&#34;, LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63756630288, loc:(*time.Location)(0x770c940)}}, Reason:&#34;ProgressDeadlineExceeded&#34;, Message:&#34;ReplicaSet \&#34;webserver-7d5989497\&#34; has timed out progressing.&#34;}}, CollisionCount:(*int32)(nil)}&#xA;occurred&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/apps.testIterativeDeployments(0xc00049d600)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:648 +0x184f&#xA;k8s.io/kubernetes/test/e2e/apps.glob..func4.8()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:122 +0x2a&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc002077500)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc002077500)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc002077500, 0x4de37a0)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;[AfterEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72&#xA;May 14 23:09:23.780: INFO: Deployment &#34;webserver&#34;:&#xA;&amp;Deployment{ObjectMeta:{webserver  deployment-2428 /apis/apps/v1/namespaces/deployment-2428/deployments/webserver 288ca9e3-ca06-49d2-b7e5-789ec435fad7 69665 23 2021-05-14 23:03:48 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd] map[deployment.kubernetes.io/revision:11] [] []  []},Spec:DeploymentSpec{Replicas:*3,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [{A 0 nil} {A 12 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc000a37ff8 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&amp;RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*2,Paused:false,ProgressDeadlineSeconds:*30,},Status:DeploymentStatus{ObservedGeneration:23,Replicas:4,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:4,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2021-05-14 23:03:48 +0000 UTC,LastTransitionTime:2021-05-14 23:03:48 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:False,Reason:ProgressDeadlineExceeded,Message:ReplicaSet &#34;webserver-7d5989497&#34; has timed out progressing.,LastUpdateTime:2021-05-14 23:04:48 +0000 UTC,LastTransitionTime:2021-05-14 23:04:48 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},}&#xA;&#xA;May 14 23:09:23.784: INFO: New ReplicaSet &#34;webserver-7d5989497&#34; of Deployment &#34;webserver&#34;:&#xA;&amp;ReplicaSet{ObjectMeta:{webserver-7d5989497  deployment-2428 /apis/apps/v1/namespaces/deployment-2428/replicasets/webserver-7d5989497 04475529-6892-423b-9b85-d84cda59114c 69385 4 2021-05-14 23:04:07 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:7d5989497] map[deployment.kubernetes.io/desired-replicas:3 deployment.kubernetes.io/max-replicas:4 deployment.kubernetes.io/revision:11 deployment.kubernetes.io/revision-history:9] [{apps/v1 Deployment webserver 288ca9e3-ca06-49d2-b7e5-789ec435fad7 0xc003e82c70 0xc003e82c71}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 7d5989497,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:7d5989497] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [{A 0 nil} {A 12 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003e82cc8 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:4,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}&#xA;May 14 23:09:23.784: INFO: All old ReplicaSets of Deployment &#34;webserver&#34;:&#xA;May 14 23:09:23.784: INFO: &amp;ReplicaSet{ObjectMeta:{webserver-7746d44bfb  deployment-2428 /apis/apps/v1/namespaces/deployment-2428/replicasets/webserver-7746d44bfb 82fc5dca-6cb5-4bf3-820b-d4a0214219ee 69284 4 2021-05-14 23:03:48 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:7746d44bfb] map[deployment.kubernetes.io/desired-replicas:4 deployment.kubernetes.io/max-replicas:5 deployment.kubernetes.io/revision:8 deployment.kubernetes.io/revision-history:2,4,6] [{apps/v1 Deployment webserver 288ca9e3-ca06-49d2-b7e5-789ec435fad7 0xc003e82ba0 0xc003e82ba1}] []  [{kube-controller-manager Update apps/v1 2021-05-14 23:04:17 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:deployment.kubernetes.io/desired-replicas&#34;:{},&#34;f:deployment.kubernetes.io/max-replicas&#34;:{},&#34;f:deployment.kubernetes.io/revision&#34;:{},&#34;f:deployment.kubernetes.io/revision-history&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;288ca9e3-ca06-49d2-b7e5-789ec435fad7\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:replicas&#34;:{},&#34;f:selector&#34;:{&#34;f:matchLabels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:template&#34;:{&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:env&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;name\&#34;:\&#34;A\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:value&#34;:{}}},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}},&#34;f:status&#34;:{&#34;f:observedGeneration&#34;:{},&#34;f:replicas&#34;:{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 7746d44bfb,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:7746d44bfb] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [{A 0 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003e82c08 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:4,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}&#xA;May 14 23:09:23.784: INFO: &amp;ReplicaSet{ObjectMeta:{webserver-dd94f59b7  deployment-2428 /apis/apps/v1/namespaces/deployment-2428/replicasets/webserver-dd94f59b7 08147cb6-49b6-41e4-a192-bf285c40ab3a 69379 9 2021-05-14 23:03:48 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:dd94f59b7] map[deployment.kubernetes.io/desired-replicas:3 deployment.kubernetes.io/max-replicas:4 deployment.kubernetes.io/revision:7 deployment.kubernetes.io/revision-history:1,3,5] [{apps/v1 Deployment webserver 288ca9e3-ca06-49d2-b7e5-789ec435fad7 0xc003e82d30 0xc003e82d31}] []  [{kube-controller-manager Update apps/v1 2021-05-14 23:03:51 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:deployment.kubernetes.io/desired-replicas&#34;:{},&#34;f:deployment.kubernetes.io/max-replicas&#34;:{},&#34;f:deployment.kubernetes.io/revision&#34;:{},&#34;f:deployment.kubernetes.io/revision-history&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;288ca9e3-ca06-49d2-b7e5-789ec435fad7\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:replicas&#34;:{},&#34;f:selector&#34;:{&#34;f:matchLabels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:template&#34;:{&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}},&#34;f:status&#34;:{&#34;f:fullyLabeledReplicas&#34;:{},&#34;f:observedGeneration&#34;:{},&#34;f:replicas&#34;:{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: dd94f59b7,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:dd94f59b7] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003e82d98 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:9,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}&#xA;May 14 23:09:23.784: INFO: &amp;ReplicaSet{ObjectMeta:{webserver-5dc9cd4855  deployment-2428 /apis/apps/v1/namespaces/deployment-2428/replicasets/webserver-5dc9cd4855 3294e033-c92d-47c6-a0cd-b35066c9d085 69376 2 2021-05-14 23:04:17 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5dc9cd4855] map[deployment.kubernetes.io/desired-replicas:3 deployment.kubernetes.io/max-replicas:4 deployment.kubernetes.io/revision:10] [{apps/v1 Deployment webserver 288ca9e3-ca06-49d2-b7e5-789ec435fad7 0xc003e82ad0 0xc003e82ad1}] []  []},Spec:ReplicaSetSpec{Replicas:*2,Selector:&amp;v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 5dc9cd4855,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5dc9cd4855] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [{A 0 nil} {A 12 nil} {A 17 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003e82b30 &lt;nil&gt; ClusterFirst map[]   &lt;nil&gt;  false false false &lt;nil&gt; &amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} []   nil default-scheduler [] []  &lt;nil&gt; nil [] &lt;nil&gt; &lt;nil&gt; &lt;nil&gt; map[] [] &lt;nil&gt;}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:2,FullyLabeledReplicas:2,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}&#xA;May 14 23:09:23.787: INFO: Pod &#34;webserver-5dc9cd4855-qk9rp&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-5dc9cd4855-qk9rp webserver-5dc9cd4855- deployment-2428 /api/v1/namespaces/deployment-2428/pods/webserver-5dc9cd4855-qk9rp 1991f603-8217-4d3b-98b2-3149585c978b 70966 0 2021-05-14 23:04:17 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5dc9cd4855] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.11&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;7e:78:1e:be:49:ec&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.11&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;7e:78:1e:be:49:ec&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-5dc9cd4855 3294e033-c92d-47c6-a0cd-b35066c9d085 0xc003e8346f 0xc003e83480}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-cmgtz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-cmgtz,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A,Value:0,ValueFrom:nil,},EnvVar{Name:A,Value:12,ValueFrom:nil,},EnvVar{Name:A,Value:17,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-cmgtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.207,PodIP:10.244.3.11,StartTime:2021-05-14 23:04:17 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.3.11,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;May 14 23:09:23.788: INFO: Pod &#34;webserver-5dc9cd4855-w9qgj&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-5dc9cd4855-w9qgj webserver-5dc9cd4855- deployment-2428 /api/v1/namespaces/deployment-2428/pods/webserver-5dc9cd4855-w9qgj 2ff5bfb8-0c08-4757-8e92-10c29817aaf5 71012 0 2021-05-14 23:04:17 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:5dc9cd4855] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.82&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;12:3e:f9:95:8c:c4&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.82&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;12:3e:f9:95:8c:c4&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-5dc9cd4855 3294e033-c92d-47c6-a0cd-b35066c9d085 0xc003e835ff 0xc003e83610}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-cmgtz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-cmgtz,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A,Value:0,ValueFrom:nil,},EnvVar{Name:A,Value:12,ValueFrom:nil,},EnvVar{Name:A,Value:17,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-cmgtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:17 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.208,PodIP:10.244.4.82,StartTime:2021-05-14 23:04:17 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.4.82,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;May 14 23:09:23.788: INFO: Pod &#34;webserver-7d5989497-2t897&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-7d5989497-2t897 webserver-7d5989497- deployment-2428 /api/v1/namespaces/deployment-2428/pods/webserver-7d5989497-2t897 754f0cb1-abda-49db-ab88-7e0d75c36679 70979 0 2021-05-14 23:04:07 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:7d5989497] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.80&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;7e:0b:c6:8a:3a:de&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.4.80&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;7e:0b:c6:8a:3a:de&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-7d5989497 04475529-6892-423b-9b85-d84cda59114c 0xc003e8378f 0xc003e837a0}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-cmgtz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-cmgtz,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A,Value:0,ValueFrom:nil,},EnvVar{Name:A,Value:12,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-cmgtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:04:07 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.208,PodIP:10.244.4.80,StartTime:2021-05-14 23:04:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.4.80,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;May 14 23:09:23.788: INFO: Pod &#34;webserver-dd94f59b7-qkqbk&#34; is not available:&#xA;&amp;Pod{ObjectMeta:{webserver-dd94f59b7-qkqbk webserver-dd94f59b7- deployment-2428 /api/v1/namespaces/deployment-2428/pods/webserver-dd94f59b7-qkqbk 593c67ba-7bb7-4bdc-8e40-9c46cc0e448e 70908 0 2021-05-14 23:03:48 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[name:httpd pod-template-hash:dd94f59b7] map[k8s.v1.cni.cncf.io/network-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.252&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;62:a5:1b:75:4b:16&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] k8s.v1.cni.cncf.io/networks-status:[{&#xA;    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;    &#34;interface&#34;: &#34;eth0&#34;,&#xA;    &#34;ips&#34;: [&#xA;        &#34;10.244.3.252&#34;&#xA;    ],&#xA;    &#34;mac&#34;: &#34;62:a5:1b:75:4b:16&#34;,&#xA;    &#34;default&#34;: true,&#xA;    &#34;dns&#34;: {}&#xA;}] kubernetes.io/psp:collectd] [{apps/v1 ReplicaSet webserver-dd94f59b7 08147cb6-49b6-41e4-a192-bf285c40ab3a 0xc003e8391f 0xc003e83930}] []  [{kube-controller-manager Update v1 2021-05-14 23:03:48 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:generateName&#34;:{},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:name&#34;:{},&#34;f:pod-template-hash&#34;:{}},&#34;f:ownerReferences&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;uid\&#34;:\&#34;08147cb6-49b6-41e4-a192-bf285c40ab3a\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:apiVersion&#34;:{},&#34;f:blockOwnerDeletion&#34;:{},&#34;f:controller&#34;:{},&#34;f:kind&#34;:{},&#34;f:name&#34;:{},&#34;f:uid&#34;:{}}}},&#34;f:spec&#34;:{&#34;f:containers&#34;:{&#34;k:{\&#34;name\&#34;:\&#34;httpd\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:image&#34;:{},&#34;f:imagePullPolicy&#34;:{},&#34;f:name&#34;:{},&#34;f:resources&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationMessagePath&#34;:{},&#34;f:terminationMessagePolicy&#34;:{}}},&#34;f:dnsPolicy&#34;:{},&#34;f:enableServiceLinks&#34;:{},&#34;f:restartPolicy&#34;:{},&#34;f:schedulerName&#34;:{},&#34;f:securityContext&#34;:{},&#34;f:terminationGracePeriodSeconds&#34;:{}}}} {multus Update v1 2021-05-14 23:03:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:k8s.v1.cni.cncf.io/network-status&#34;:{},&#34;f:k8s.v1.cni.cncf.io/networks-status&#34;:{}}}}} {kubelet Update v1 2021-05-14 23:03:59 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;ContainersReady\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Initialized\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastProbeTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:containerStatuses&#34;:{},&#34;f:hostIP&#34;:{},&#34;f:podIP&#34;:{},&#34;f:podIPs&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;ip\&#34;:\&#34;10.244.3.252\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:ip&#34;:{}}},&#34;f:startTime&#34;:{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-cmgtz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&amp;SecretVolumeSource{SecretName:default-token-cmgtz,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-cmgtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&amp;SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:node1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&amp;PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:03:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:03:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:03:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-05-14 23:03:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.10.190.207,PodIP:10.244.3.252,StartTime:2021-05-14 23:03:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&amp;ContainerStateWaiting{Reason:ImagePullBackOff,Message:Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.3.252,},},EphemeralContainerStatuses:[]ContainerStatus{},},}&#xA;[AfterEach] [sig-apps] Deployment&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175&#xA;STEP: Collecting events from namespace &#34;deployment-2428&#34;.&#xA;STEP: Found 159 events.&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5dc9cd4855-qk9rp: { } Scheduled: Successfully assigned deployment-2428/webserver-5dc9cd4855-qk9rp to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-5dc9cd4855-w9qgj: { } Scheduled: Successfully assigned deployment-2428/webserver-5dc9cd4855-w9qgj to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7746d44bfb-4s4nr: { } Scheduled: Successfully assigned deployment-2428/webserver-7746d44bfb-4s4nr to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7746d44bfb-bq496: { } Scheduled: Successfully assigned deployment-2428/webserver-7746d44bfb-bq496 to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7746d44bfb-cqwzl: { } Scheduled: Successfully assigned deployment-2428/webserver-7746d44bfb-cqwzl to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7746d44bfb-d6p9c: { } Scheduled: Successfully assigned deployment-2428/webserver-7746d44bfb-d6p9c to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7746d44bfb-h5wnd: { } Scheduled: Successfully assigned deployment-2428/webserver-7746d44bfb-h5wnd to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7746d44bfb-pnq8l: { } Scheduled: Successfully assigned deployment-2428/webserver-7746d44bfb-pnq8l to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7d5989497-2t897: { } Scheduled: Successfully assigned deployment-2428/webserver-7d5989497-2t897 to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7d5989497-d2rzc: { } Scheduled: Successfully assigned deployment-2428/webserver-7d5989497-d2rzc to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-7d5989497-ptzf6: { } Scheduled: Successfully assigned deployment-2428/webserver-7d5989497-ptzf6 to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-8tmzq: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-8tmzq to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-d8b9m: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-d8b9m to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-fcgzx: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-fcgzx to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-g8jb5: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-g8jb5 to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-hw67g: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-hw67g to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-j6hlt: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-j6hlt to node2&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-lxk7s: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-lxk7s to node1&#xA;May 14 23:09:23.797: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for webserver-dd94f59b7-qkqbk: { } Scheduled: Successfully assigned deployment-2428/webserver-dd94f59b7-qkqbk to node1&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-dd94f59b7 to 7&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-7746d44bfb to 2&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-7746d44bfb to 4&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-dd94f59b7 to 6&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-dd94f59b7 to 6&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-dd94f59b7 to 8&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7746d44bfb-bq496&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7746d44bfb-h5wnd&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7746d44bfb-4s4nr&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7746d44bfb-d6p9c&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-lxk7s&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-fcgzx&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-g8jb5&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-8tmzq&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-qkqbk&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:48 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-d8b9m&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:50 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7746d44bfb-pnq8l&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:50 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7746d44bfb-cqwzl&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:50 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-hw67g&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:50 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulCreate: Created pod: webserver-dd94f59b7-j6hlt&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:51 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 1&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:51 +0000 UTC - event for webserver-dd94f59b7-lxk7s: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:52 +0000 UTC - event for webserver-7746d44bfb-4s4nr: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:52 +0000 UTC - event for webserver-7746d44bfb-4s4nr: {multus } AddedInterface: Add eth0 [10.244.3.251/24]&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:52 +0000 UTC - event for webserver-dd94f59b7-lxk7s: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:52 +0000 UTC - event for webserver-dd94f59b7-lxk7s: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:53 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 2&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:53 +0000 UTC - event for webserver-7746d44bfb-4s4nr: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:53 +0000 UTC - event for webserver-7746d44bfb-4s4nr: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:53 +0000 UTC - event for webserver-dd94f59b7-g8jb5: {kubelet node2} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to start sandbox container for pod &#34;webserver-dd94f59b7-g8jb5&#34;: Error response from daemon: OCI runtime create failed: container_linux.go:349: starting container process caused &#34;process_linux.go:365: sending config to init process caused \&#34;write init-p: broken pipe\&#34;&#34;: unknown&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:53 +0000 UTC - event for webserver-dd94f59b7-qkqbk: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:53 +0000 UTC - event for webserver-dd94f59b7-qkqbk: {multus } AddedInterface: Add eth0 [10.244.3.252/24]&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:54 +0000 UTC - event for webserver-dd94f59b7-qkqbk: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:54 +0000 UTC - event for webserver-dd94f59b7-qkqbk: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:55 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 3&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:55 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-dd94f59b7 to 5&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:55 +0000 UTC - event for webserver-7746d44bfb-h5wnd: {kubelet node1} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to start sandbox container for pod &#34;webserver-7746d44bfb-h5wnd&#34;: Error response from daemon: OCI runtime create failed: container_linux.go:349: starting container process caused &#34;process_linux.go:365: sending config to init process caused \&#34;write init-p: broken pipe\&#34;&#34;: unknown&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:55 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-j6hlt&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:55 +0000 UTC - event for webserver-dd94f59b7-d8b9m: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:55 +0000 UTC - event for webserver-dd94f59b7-d8b9m: {multus } AddedInterface: Add eth0 [10.244.4.70/24]&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:55 +0000 UTC - event for webserver-dd94f59b7-fcgzx: {multus } AddedInterface: Add eth0 [10.244.3.253/24]&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:56 +0000 UTC - event for webserver-dd94f59b7-8tmzq: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:56 +0000 UTC - event for webserver-dd94f59b7-8tmzq: {multus } AddedInterface: Add eth0 [10.244.4.72/24]&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:56 +0000 UTC - event for webserver-dd94f59b7-d8b9m: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:56 +0000 UTC - event for webserver-dd94f59b7-d8b9m: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:57 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 4&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:57 +0000 UTC - event for webserver-7746d44bfb-bq496: {multus } AddedInterface: Add eth0 [10.244.4.73/24]&#xA;May 14 23:09:23.797: INFO: At 2021-05-14 23:03:57 +0000 UTC - event for webserver-7746d44bfb-bq496: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:57 +0000 UTC - event for webserver-dd94f59b7-fcgzx: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:58 +0000 UTC - event for webserver-dd94f59b7-8tmzq: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:58 +0000 UTC - event for webserver-dd94f59b7-8tmzq: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:58 +0000 UTC - event for webserver-dd94f59b7-fcgzx: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:58 +0000 UTC - event for webserver-dd94f59b7-fcgzx: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-7746d44bfb-4s4nr: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-7746d44bfb-4s4nr: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-7746d44bfb-bq496: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-7746d44bfb-bq496: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-dd94f59b7-fcgzx: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-dd94f59b7-fcgzx: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-dd94f59b7-j6hlt: {kubelet node2} FailedCreatePodSandBox: Failed to create pod sandbox: rpc error: code = Unknown desc = failed to start sandbox container for pod &#34;webserver-dd94f59b7-j6hlt&#34;: Error response from daemon: OCI runtime create failed: container_linux.go:349: starting container process caused &#34;process_linux.go:449: container init caused \&#34;process_linux.go:432: running prestart hook 0 caused \\\&#34;error running hook: exit status 1, stdout: , stderr: time=\\\\\\\&#34;2021-05-14T23:03:58Z\\\\\\\&#34; level=fatal msg=\\\\\\\&#34;no such file or directory\\\\\\\&#34;\\\\n\\\&#34;\&#34;&#34;: unknown&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-dd94f59b7-qkqbk: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:03:59 +0000 UTC - event for webserver-dd94f59b7-qkqbk: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 5&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver-7746d44bfb-cqwzl: {multus } AddedInterface: Add eth0 [10.244.4.74/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver-7746d44bfb-cqwzl: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver-7746d44bfb-pnq8l: {multus } AddedInterface: Add eth0 [10.244.3.254/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver-dd94f59b7-8tmzq: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver-dd94f59b7-8tmzq: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver-dd94f59b7-d8b9m: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:00 +0000 UTC - event for webserver-dd94f59b7-d8b9m: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-7746d44bfb-bq496: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-7746d44bfb-bq496: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-7746d44bfb-cqwzl: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-7746d44bfb-cqwzl: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-7746d44bfb-cqwzl: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-7746d44bfb-cqwzl: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-7746d44bfb-pnq8l: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-dd94f59b7-hw67g: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:01 +0000 UTC - event for webserver-dd94f59b7-hw67g: {multus } AddedInterface: Add eth0 [10.244.3.2/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:03 +0000 UTC - event for webserver-7746d44bfb-pnq8l: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:03 +0000 UTC - event for webserver-7746d44bfb-pnq8l: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:04 +0000 UTC - event for webserver-7746d44bfb-pnq8l: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:04 +0000 UTC - event for webserver-7746d44bfb-pnq8l: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:04 +0000 UTC - event for webserver-dd94f59b7-hw67g: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:04 +0000 UTC - event for webserver-dd94f59b7-hw67g: {kubelet node1} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:04 +0000 UTC - event for webserver-dd94f59b7-hw67g: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:05 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 6&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:06 +0000 UTC - event for webserver-dd94f59b7-hw67g: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:06 +0000 UTC - event for webserver-dd94f59b7-hw67g: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:06 +0000 UTC - event for webserver-dd94f59b7-hw67g: {multus } AddedInterface: Add eth0 [10.244.3.9/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled up replica set webserver-7d5989497 to 3&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: Scaled down replica set webserver-7746d44bfb to 1&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-7746d44bfb-pnq8l&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-7746d44bfb-4s4nr&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-7746d44bfb-cqwzl&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver-7d5989497: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7d5989497-2t897&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver-7d5989497: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7d5989497-ptzf6&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:07 +0000 UTC - event for webserver-7d5989497: {replicaset-controller } SuccessfulCreate: Created pod: webserver-7d5989497-d2rzc&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:09 +0000 UTC - event for webserver: {deployment-controller } ScalingReplicaSet: (combined from similar events): Scaled down replica set webserver-7d5989497 to 1&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:09 +0000 UTC - event for webserver-7d5989497-d2rzc: {multus } AddedInterface: Add eth0 [10.244.3.10/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:09 +0000 UTC - event for webserver-7d5989497-d2rzc: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:09 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-d8b9m&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:10 +0000 UTC - event for webserver-7d5989497-d2rzc: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:10 +0000 UTC - event for webserver-7d5989497-d2rzc: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:10 +0000 UTC - event for webserver-7d5989497-d2rzc: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:10 +0000 UTC - event for webserver-7d5989497-d2rzc: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:11 +0000 UTC - event for webserver-7d5989497-2t897: {multus } AddedInterface: Add eth0 [10.244.4.80/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:11 +0000 UTC - event for webserver-7d5989497-2t897: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:12 +0000 UTC - event for webserver-7d5989497-ptzf6: {multus } AddedInterface: Add eth0 [10.244.4.81/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:12 +0000 UTC - event for webserver-7d5989497-ptzf6: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:12 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-hw67g&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:14 +0000 UTC - event for webserver-7d5989497-2t897: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:14 +0000 UTC - event for webserver-7d5989497-2t897: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:14 +0000 UTC - event for webserver-7d5989497-2t897: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:14 +0000 UTC - event for webserver-7d5989497-2t897: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:16 +0000 UTC - event for webserver-7d5989497-ptzf6: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:16 +0000 UTC - event for webserver-7d5989497-ptzf6: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:16 +0000 UTC - event for webserver-7d5989497-ptzf6: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:16 +0000 UTC - event for webserver-7d5989497-ptzf6: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:17 +0000 UTC - event for webserver-5dc9cd4855: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5dc9cd4855-w9qgj&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:17 +0000 UTC - event for webserver-5dc9cd4855: {replicaset-controller } SuccessfulCreate: Created pod: webserver-5dc9cd4855-qk9rp&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:17 +0000 UTC - event for webserver-7746d44bfb: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-7746d44bfb-bq496&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:17 +0000 UTC - event for webserver-7d5989497: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-7d5989497-ptzf6&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:17 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-fcgzx&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:17 +0000 UTC - event for webserver-dd94f59b7: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-dd94f59b7-8tmzq&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:18 +0000 UTC - event for webserver-5dc9cd4855-qk9rp: {multus } AddedInterface: Add eth0 [10.244.3.11/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:18 +0000 UTC - event for webserver-5dc9cd4855-qk9rp: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:19 +0000 UTC - event for webserver: {deployment-controller } DeploymentRollback: Rolled back deployment &#34;webserver&#34; to revision 9&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:19 +0000 UTC - event for webserver-5dc9cd4855-qk9rp: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:19 +0000 UTC - event for webserver-5dc9cd4855-qk9rp: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:19 +0000 UTC - event for webserver-5dc9cd4855-w9qgj: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:19 +0000 UTC - event for webserver-5dc9cd4855-w9qgj: {multus } AddedInterface: Add eth0 [10.244.4.82/24]&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:20 +0000 UTC - event for webserver-5dc9cd4855-qk9rp: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:20 +0000 UTC - event for webserver-5dc9cd4855-qk9rp: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:20 +0000 UTC - event for webserver-5dc9cd4855-w9qgj: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/httpd:2.4.38-alpine&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:20 +0000 UTC - event for webserver-5dc9cd4855-w9qgj: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:20 +0000 UTC - event for webserver-5dc9cd4855-w9qgj: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:20 +0000 UTC - event for webserver-5dc9cd4855-w9qgj: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/httpd:2.4.38-alpine&#34;&#xA;May 14 23:09:23.798: INFO: At 2021-05-14 23:04:21 +0000 UTC - event for webserver-7d5989497: {replicaset-controller } SuccessfulDelete: Deleted pod: webserver-7d5989497-d2rzc&#xA;May 14 23:09:23.799: INFO: At 2021-05-14 23:05:53 +0000 UTC - event for webserver-7746d44bfb-d6p9c: {kubelet node2} FailedMount: Unable to attach or mount volumes: unmounted volumes=[default-token-cmgtz], unattached volumes=[default-token-cmgtz]: timed out waiting for the condition&#xA;May 14 23:09:23.802: INFO: POD                         NODE   PHASE    GRACE  CONDITIONS&#xA;May 14 23:09:23.802: INFO: webserver-5dc9cd4855-qk9rp  node1  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC  }]&#xA;May 14 23:09:23.802: INFO: webserver-5dc9cd4855-w9qgj  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:17 +0000 UTC  }]&#xA;May 14 23:09:23.802: INFO: webserver-7d5989497-2t897   node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:07 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:07 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:07 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:04:07 +0000 UTC  }]&#xA;May 14 23:09:23.802: INFO: webserver-dd94f59b7-qkqbk   node1  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:03:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:03:48 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:03:48 +0000 UTC ContainersNotReady containers with unready status: [httpd]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-14 23:03:48 +0000 UTC  }]&#xA;May 14 23:09:23.802: INFO: &#xA;May 14 23:09:23.806: INFO: &#xA;Logging node info for node master1&#xA;May 14 23:09:23.809: INFO: Node Info: &amp;Node{ObjectMeta:{master1   /api/v1/nodes/master1 8e20012e-a811-456d-9add-2ea316e23700 71683 0 2021-05-14 19:56:35 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;a6:a3:7b:a0:c9:75&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-14 19:56:34 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-14 19:56:37 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-14 19:59:03 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-14 19:59:09 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {nfd-master Update v1 2021-05-14 20:06:00 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-14 20:02:00 +0000 UTC,LastTransitionTime:2021-05-14 20:02:00 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:20 +0000 UTC,LastTransitionTime:2021-05-14 19:56:32 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:20 +0000 UTC,LastTransitionTime:2021-05-14 19:56:32 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:20 +0000 UTC,LastTransitionTime:2021-05-14 19:56:32 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-14 23:09:20 +0000 UTC,LastTransitionTime:2021-05-14 19:59:09 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:a88b162033bc4931ba0342c7f78a28b9,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:ba5ed4e5-a8ef-4986-946f-e7e2d91395d2,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726663003,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:e3157cfba16d361ffec06306dd0154c7dca1931cbc4569e3c5822e30e311948b tas-controller:latest localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[nginx@sha256:a97eb9ecc708c8aa715ccfb5e9338f5456e4b65575daf304f108301f3b497314 nginx:1.19.2-alpine],SizeBytes:22052669,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:a43c7fdd150533238a300ad84ac906e551111f9b57273afcb8781ee675fd23b3 tas-extender:latest localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 14 23:09:23.809: INFO: &#xA;Logging kubelet events for node master1&#xA;May 14 23:09:23.813: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May 14 23:09:23.834: INFO: kube-proxy-v2c76 started at 2021-05-14 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 14 23:09:23.834: INFO: kube-multus-ds-amd64-m54v2 started at 2021-05-14 19:59:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 14 23:09:23.834: INFO: coredns-7677f9bb54-mlcmv started at 2021-05-14 19:59:27 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May 14 23:09:23.834: INFO: docker-registry-docker-registry-56cbc7bc58-bjc5h started at 2021-05-14 20:02:43 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: node-feature-discovery-controller-5bf5c49849-27v77 started at 2021-05-14 20:05:52 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: kube-apiserver-master1 started at 2021-05-14 19:57:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: kube-flannel-cx7s6 started at 2021-05-14 19:58:58 +0000 UTC (1+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 14 23:09:23.834: INFO: node-exporter-nvrxr started at 2021-05-14 20:09:56 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: kube-scheduler-master1 started at 2021-05-14 19:57:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;May 14 23:09:23.834: INFO: kube-controller-manager-master1 started at 2021-05-14 20:01:22 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.834: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;W0514 23:09:23.848093      22 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 14 23:09:23.872: INFO: &#xA;Latency metrics for node master1&#xA;May 14 23:09:23.872: INFO: &#xA;Logging node info for node master2&#xA;May 14 23:09:23.875: INFO: Node Info: &amp;Node{ObjectMeta:{master2   /api/v1/nodes/master2 e8f1881f-5ded-4c6c-b7e6-eb354b7962e2 71682 0 2021-05-14 19:57:04 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;0a:97:9a:eb:9d:a8&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-14 19:57:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-14 19:57:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-14 19:59:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-14 19:59:09 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-14 20:02:07 +0000 UTC,LastTransitionTime:2021-05-14 20:02:07 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:19 +0000 UTC,LastTransitionTime:2021-05-14 19:57:04 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:19 +0000 UTC,LastTransitionTime:2021-05-14 19:57:04 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:19 +0000 UTC,LastTransitionTime:2021-05-14 19:57:04 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-14 23:09:19 +0000 UTC,LastTransitionTime:2021-05-14 19:59:09 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:14c4cdd0613041bb923c5f9b84e0bcde,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:9bdca68c-a5fc-48f7-b392-63d2c04d224d,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726663003,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 14 23:09:23.875: INFO: &#xA;Logging kubelet events for node master2&#xA;May 14 23:09:23.878: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May 14 23:09:23.885: INFO: dns-autoscaler-5b7b5c9b6f-fgzqp started at 2021-05-14 19:59:30 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Container autoscaler ready: true, restart count 2&#xA;May 14 23:09:23.885: INFO: node-exporter-gjrtc started at 2021-05-14 20:09:56 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:23.885: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 14 23:09:23.885: INFO: kube-apiserver-master2 started at 2021-05-14 20:05:10 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 14 23:09:23.885: INFO: kube-controller-manager-master2 started at 2021-05-14 20:01:22 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 14 23:09:23.885: INFO: kube-scheduler-master2 started at 2021-05-14 19:57:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;May 14 23:09:23.885: INFO: kube-proxy-qcgpm started at 2021-05-14 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 14 23:09:23.885: INFO: kube-flannel-fc4sf started at 2021-05-14 19:58:58 +0000 UTC (1+1 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 14 23:09:23.885: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 14 23:09:23.885: INFO: kube-multus-ds-amd64-bt5kr started at 2021-05-14 19:59:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.885: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;W0514 23:09:23.898861      22 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 14 23:09:23.922: INFO: &#xA;Latency metrics for node master2&#xA;May 14 23:09:23.922: INFO: &#xA;Logging node info for node master3&#xA;May 14 23:09:23.925: INFO: Node Info: &amp;Node{ObjectMeta:{master3   /api/v1/nodes/master3 29fd0a5d-1350-4e28-a4cb-b26dd82cd397 71674 0 2021-05-14 19:57:14 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;ae:27:37:b7:ad:a5&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-14 19:57:14 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-14 19:57:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-14 19:59:03 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-14 19:59:09 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-14 20:01:03 +0000 UTC,LastTransitionTime:2021-05-14 20:01:03 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:17 +0000 UTC,LastTransitionTime:2021-05-14 19:57:14 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:17 +0000 UTC,LastTransitionTime:2021-05-14 19:57:14 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:17 +0000 UTC,LastTransitionTime:2021-05-14 19:57:14 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-14 23:09:17 +0000 UTC,LastTransitionTime:2021-05-14 20:00:41 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:f3fe601830d34e59967ed389af552f25,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:56dd60e2-98fe-4d87-81d9-95db820e7426,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726663003,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 14 23:09:23.925: INFO: &#xA;Logging kubelet events for node master3&#xA;May 14 23:09:23.927: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May 14 23:09:23.935: INFO: kube-flannel-cl8jf started at 2021-05-14 19:58:58 +0000 UTC (1+1 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 14 23:09:23.935: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 14 23:09:23.935: INFO: kube-multus-ds-amd64-hp6bp started at 2021-05-14 19:59:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 14 23:09:23.935: INFO: coredns-7677f9bb54-qhzrv started at 2021-05-14 19:59:32 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Container coredns ready: true, restart count 2&#xA;May 14 23:09:23.935: INFO: node-exporter-4cgbq started at 2021-05-14 20:09:56 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:23.935: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 14 23:09:23.935: INFO: kube-controller-manager-master3 started at 2021-05-14 20:00:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Container kube-controller-manager ready: true, restart count 3&#xA;May 14 23:09:23.935: INFO: kube-scheduler-master3 started at 2021-05-14 19:57:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Container kube-scheduler ready: true, restart count 3&#xA;May 14 23:09:23.935: INFO: kube-apiserver-master3 started at 2021-05-14 19:57:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 14 23:09:23.935: INFO: kube-proxy-2crs2 started at 2021-05-14 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:23.935: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;W0514 23:09:23.947174      22 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 14 23:09:23.975: INFO: &#xA;Latency metrics for node master3&#xA;May 14 23:09:23.975: INFO: &#xA;Logging node info for node node1&#xA;May 14 23:09:23.978: INFO: Node Info: &amp;Node{ObjectMeta:{node1   /api/v1/nodes/node1 5e4c0fba-b5fa-4177-b834-f3e04c846ff3 71694 0 2021-05-14 19:58:21 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;ba:ee:c6:a6:52:03&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-14 19:58:21 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-14 19:58:22 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-14 19:59:03 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-14 20:06:02 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-14 20:08:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-14 22:28:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-14 20:00:44 +0000 UTC,LastTransitionTime:2021-05-14 20:00:44 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 19:58:21 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 19:58:21 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 19:58:21 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 20:00:41 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:4b96d01fdbcb4fadb4a59fca2e1bf368,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:93c238b3-1895-423c-a1aa-193fbcf8b55f,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:869f7b3516c269b43448f1227c57d536e8a4cf723eeef3b5f8b8e224ecbcfd8e localhost:30500/barometer-collectd:stable],SizeBytes:1464261626,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1002487751,},ContainerImage{Names:[opnfv/barometer-collectd@sha256:ed5c574f653e2a39e784ff322033a2319aafde7366c803a88f20f7a2a8bc1efb opnfv/barometer-collectd:stable],SizeBytes:825413035,},ContainerImage{Names:[localhost:30500/cmk@sha256:f759b012c29126f880575ac543d09301d45f0b2b9d0f5329849ea40e65017dde cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726663003,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:1636899c10870ab66c48d960a9df620f4f9e86a0c72fbacf36032d27404e7e6c golang:alpine3.12],SizeBytes:301156062,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[grafana/grafana@sha256:89304bc2335f4976618548d7b93d165ed67369d3a051d2f627fc4e0aa3d0aff1 grafana/grafana:7.1.0],SizeBytes:179601493,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:d4ba4dd1a9ebb90916d0bfed3c204adcb118ed24546bf8dd2e6b30fc0fd2009e quay.io/prometheus/prometheus:v2.20.0],SizeBytes:144886595,},ContainerImage{Names:[nginx@sha256:df13abe416e37eb3db4722840dd479b00ba193ac6606e7902331dcea50f4f1f2 nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/sample-apiserver@sha256:ff02aacd9766d597883fabafc7ad604c719a57611db1bcc1564c69a45b000a55 gcr.io/kubernetes-e2e-test-images/sample-apiserver:1.17],SizeBytes:60684726,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter-amd64@sha256:b63dc612e3cb73f79d2401a4516f794f9f0a83002600ca72e675e41baecff437 directxman12/k8s-prometheus-adapter-amd64:v0.6.0],SizeBytes:53267842,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:948a93bc3803d61dd66ab49f99d4cc657e87273682aec7dd5955a000fd17a7e5 nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392645,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:9d07c391aeb1a9d02eb4343c113ed01825227c70c32b3cae861711f90191b0fd quay.io/coreos/kube-rbac-proxy:v0.4.1],SizeBytes:41317870,},ContainerImage{Names:[quay.io/coreos/prometheus-operator@sha256:a54e806fb27d2fb0251da4f3b2a3bb5320759af63a54a755788304775f2384a7 quay.io/coreos/prometheus-operator:v0.40.0],SizeBytes:38238457,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[quay.io/coreos/prometheus-config-reloader@sha256:c679a143b24b7731ad1577a9865aa3805426cbf1b25e30807b951dff68466ffd quay.io/coreos/prometheus-config-reloader:v0.40.0],SizeBytes:10131705,},ContainerImage{Names:[jimmidyson/configmap-reload@sha256:d107c7a235c266273b1c3502a391fec374430e5625539403d0de797fa9c556a2 jimmidyson/configmap-reload:v0.3.0],SizeBytes:9700438,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:36553b10a4947067b9fbb7d532951066293a68eae893beba1d9235f7d11a20ad alpine:3.12],SizeBytes:5581415,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 14 23:09:23.979: INFO: &#xA;Logging kubelet events for node node1&#xA;May 14 23:09:23.982: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May 14 23:09:24.002: INFO: nginx-proxy-node1 started at 2021-05-14 20:05:10 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 14 23:09:24.002: INFO: cmk-init-discover-node1-g6c6q started at 2021-05-14 20:08:21 +0000 UTC (0+3 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: all-succeed-sxvbk started at 2021-05-14 23:03:31 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: webserver-dd94f59b7-qkqbk started at 2021-05-14 23:03:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: node-feature-discovery-worker-nvg9t started at 2021-05-14 20:05:42 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: node-exporter-r62rn started at 2021-05-14 20:09:56 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: kube-flannel-hj8sj started at 2021-05-14 19:58:58 +0000 UTC (1+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 14 23:09:24.002: INFO: fail-once-non-local-4hdtk started at 2021-05-14 23:03:27 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: test-orphan-deployment-dd94f59b7-rvwbp started at 2021-05-14 23:04:23 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: kube-multus-ds-amd64-jhf4c started at 2021-05-14 19:59:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 14 23:09:24.002: INFO: test-new-deployment-dd94f59b7-8wmzh started at 2021-05-14 23:09:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: all-succeed-phzpb started at 2021-05-14 23:03:31 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: kubernetes-metrics-scraper-678c97765c-pqnd6 started at 2021-05-14 19:59:31 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May 14 23:09:24.002: INFO: cmk-9nqjc started at 2021-05-14 20:09:03 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: prometheus-k8s-0 started at 2021-05-14 20:09:58 +0000 UTC (0+5 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May 14 23:09:24.002: INFO: &#x9;Container prometheus-config-reloader ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container rules-configmap-reloader ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: prometheus-operator-5bb8cb9d8f-4lq8j started at 2021-05-14 20:09:49 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: backofflimit-j757c started at 2021-05-14 23:03:46 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: kube-proxy-l7697 started at 2021-05-14 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 14 23:09:24.002: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-bw262 started at 2021-05-14 20:06:38 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: collectd-qhp77 started at 2021-05-14 20:15:36 +0000 UTC (0+3 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:24.002: INFO: fail-once-non-local-nqqt6 started at 2021-05-14 23:03:27 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.002: INFO: webserver-5dc9cd4855-qk9rp started at 2021-05-14 23:04:17 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.002: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;W0514 23:09:24.015038      22 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 14 23:09:24.046: INFO: &#xA;Latency metrics for node node1&#xA;May 14 23:09:24.046: INFO: &#xA;Logging node info for node node2&#xA;May 14 23:09:24.049: INFO: Node Info: &amp;Node{ObjectMeta:{node2   /api/v1/nodes/node2 0bae98dc-2ebc-4849-b99e-7780a3bea71e 71693 0 2021-05-14 19:58:22 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;c6:18:ed:95:bb:1a&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-14 19:58:22 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-14 19:58:22 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-14 19:59:03 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-14 20:06:00 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-14 20:08:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {e2e.test Update v1 2021-05-14 22:19:21 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:example.com/fakecpu&#34;:{}}}}} {kubelet Update v1 2021-05-14 22:28:17 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:example.com/fakecpu&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-14 20:01:27 +0000 UTC,LastTransitionTime:2021-05-14 20:01:27 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 19:58:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 19:58:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 19:58:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-14 23:09:22 +0000 UTC,LastTransitionTime:2021-05-14 19:59:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:a3f22fbf9e534ba1819f7a549414a8a6,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:166b6e45-ba8b-4b89-80b0-befc9a0152b8,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:869f7b3516c269b43448f1227c57d536e8a4cf723eeef3b5f8b8e224ecbcfd8e localhost:30500/barometer-collectd:stable],SizeBytes:1464261626,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:726663003,},ContainerImage{Names:[localhost:30500/cmk@sha256:f759b012c29126f880575ac543d09301d45f0b2b9d0f5329849ea40e65017dde localhost:30500/cmk:v1.5.1],SizeBytes:726663003,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[nginx@sha256:df13abe416e37eb3db4722840dd479b00ba193ac6606e7902331dcea50f4f1f2 nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:948a93bc3803d61dd66ab49f99d4cc657e87273682aec7dd5955a000fd17a7e5 localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392645,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:e3157cfba16d361ffec06306dd0154c7dca1931cbc4569e3c5822e30e311948b localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:a43c7fdd150533238a300ad84ac906e551111f9b57273afcb8781ee675fd23b3 localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[gcr.io/authenticated-image-pulling/alpine@sha256:7ff177862cb50c602bfe81f805969412e619c054a2bbead977d0c276988aa4a0 gcr.io/authenticated-image-pulling/alpine:3.7],SizeBytes:4206620,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 14 23:09:24.049: INFO: &#xA;Logging kubelet events for node node2&#xA;May 14 23:09:24.052: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May 14 23:09:24.065: INFO: cmk-webhook-6c9d5f8578-pjgxh started at 2021-05-14 20:09:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.065: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May 14 23:09:24.065: INFO: cmk-init-discover-node2-j75ff started at 2021-05-14 20:08:41 +0000 UTC (0+3 container statuses recorded)&#xA;May 14 23:09:24.065: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 14 23:09:24.065: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: ss2-0 started at 2021-05-14 23:03:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container webserver ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: webserver-7d5989497-2t897 started at 2021-05-14 23:04:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: kube-proxy-t524z started at 2021-05-14 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 14 23:09:24.066: INFO: kube-multus-ds-amd64-n7cb2 started at 2021-05-14 19:59:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 14 23:09:24.066: INFO: kubernetes-dashboard-86c6f9df5b-ndntg started at 2021-05-14 19:59:31 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 2&#xA;May 14 23:09:24.066: INFO: nginx-proxy-node2 started at 2021-05-14 20:05:10 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 14 23:09:24.066: INFO: webserver-5dc9cd4855-w9qgj started at 2021-05-14 23:04:17 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container httpd ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: kube-flannel-rqcwp started at 2021-05-14 19:58:58 +0000 UTC (1+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Init container install-cni ready: true, restart count 1&#xA;May 14 23:09:24.066: INFO: &#x9;Container kube-flannel ready: true, restart count 4&#xA;May 14 23:09:24.066: INFO: node-exporter-rnd5f started at 2021-05-14 20:09:56 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: tas-telemetry-aware-scheduling-575ccbc9d4-l5hlq started at 2021-05-14 20:12:48 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container tas-controller ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: collectd-xzrgs started at 2021-05-14 20:15:36 +0000 UTC (0+3 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: all-pods-removed-d9pm9 started at 2021-05-14 23:08:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-2c2pw started at 2021-05-14 20:06:38 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: concurrent-1621033740-dgvcx started at 2021-05-14 23:09:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: all-pods-removed-n9vx7 started at 2021-05-14 23:08:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 14 23:09:24.066: INFO: node-feature-discovery-worker-76m6b started at 2021-05-14 20:05:42 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: cmk-d2qwf started at 2021-05-14 20:09:04 +0000 UTC (0+2 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 14 23:09:24.066: INFO: failed-jobs-history-limit-1621033740-hmrq5 started at 2021-05-14 23:09:07 +0000 UTC (0+1 container statuses recorded)&#xA;May 14 23:09:24.066: INFO: &#x9;Container c ready: false, restart count 0&#xA;W0514 23:09:24.080742      22 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 14 23:09:24.126: INFO: &#xA;Latency metrics for node node2&#xA;May 14 23:09:24.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;deployment-2428&#34; for this suite.&#xA;</system-out>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Docker Containers should be able to override the image&#39;s default command (docker entrypoint) [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Simple pod should return command exit codes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable[Feature:VolumeSnapshotDataSource] volume snapshot controller  should check snapshot fields, check restore correctly works after modifying source data, check deletion" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a persistent volume claim. [sig-storage]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Downward API should provide host IP and pod IP as an env var if pod uses host network [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Services should function for node-Service: http" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Generic Ephemeral-volume (default fs) [Feature:GenericEphemeralVolume] (late-binding)] ephemeral should support two pods which share the same volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  StatefulSet with pod affinity [Slow] should use volumes spread across nodes when pod management is parallel and pod has anti-affinity" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] ServiceAccounts should support InClusterConfig with token rotation [Slow] [Feature:TokenRequestProjection]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:NEG] should be able to switch between IG and NEG modes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking should provide Internet connection for containers [Feature:Networking-IPv6][Experimental][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
  </testsuite>