<?xml version="1.0" encoding="UTF-8"?>
  <testsuite name="Kubernetes e2e suite" tests="23" failures="2" errors="0" time="1448.51">
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="8.099431659"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Probing container should be restarted with a docker exec liveness probe with timeout " classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Network Partition [Disruptive] [Slow] [k8s.io] [ReplicationController] should eagerly create replacement pod during network partition when termination grace is non-zero" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be rejected when no endpoints exist" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="26.11481497"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be able to change the type and ports of a service [Slow] [DisabledForLargeClusters]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] [Serial] Volume metrics should create volume metrics in Volume Manager" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="8.116752391"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] volume on tmpfs should have the correct mode using FSGroup" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] StatefulSet [k8s.io] Deploy clustered applications [Feature:StatefulSet] [Slow] should creating a working mysql cluster" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DisruptionController evictions: enough pods, replicaSet, percentage =&gt; should allow an eviction" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:Ingress] multicluster ingress should get instance group annotation" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] [Serial] Volume metrics should create metrics for total number of volumes in A/D Controller" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] MetricsGrabber should grab all metrics from a ControllerManager." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI Volume expansion should expand volume without restarting pod if nodeExpansion=off" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PVC Protection Verify &#34;immediate&#34; deletion of a PVC that is not in active use by a pod" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV, based on multiple zones specified in storage class " classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [Feature:Example] [k8s.io] Liveness liveness pods should be automatically restarted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] kubelet [k8s.io] [sig-node] host cleanup with volume mounts [sig-storage][HostCleanup][Flaky] Host cleanup after disrupting NFS volume [NFS] after stopping the nfs-server and deleting the (sleeping) client pod, the NFS mount and the pod&#39;s UID directory should be removed." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Simple pod should support port-forward" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota&#39;s priority class scope (cpu, memory quota set) against a pod with same priority class." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="4.055215619"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI NodeStage error cases [Slow] should retry NodeStage after NodeStage ephemeral error" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should have session affinity work for LoadBalancer service with ESIPP on [Slow] [DisabledForLargeClusters] [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] [Feature:NodeAuthorizer] Getting a secret for a workload the node has access to should succeed" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="6.067236643"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected secret Should fail non-optional pod creation due to the key in the secret object does not exist [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="10.085625942"></testcase>
      <testcase name="[sig-network] NoSNAT [Feature:NoSNAT] [Slow] Should be able to send traffic between Pods without SNAT" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PV Protection Verify &#34;immediate&#34; deletion of a PV that is not bound to a PVC" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DaemonRestart [Disruptive] Kubelet should not restart containers across restart" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DisruptionController evictions: maxUnavailable deny evictions, integer =&gt; should not allow an eviction [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-bindmounted] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:NEG] should create NEGs for all ports with the Ingress annotation, and NEGs for the standalone annotation otherwise" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota&#39;s priority class scope (quota set to pod count: 1) against a pod with different priority class (ScopeSelectorOpNotIn)." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: gce-localssd-scsi-fs] [Serial] Set fsGroup for local volume should set fsGroup for one pod [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="26.116625042"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance]" classname="Kubernetes e2e suite" time="0.083593476"></testcase>
      <testcase name="[sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should allow privilege escalation when not explicitly set and uid != 0 [LinuxOnly] [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Provisioning On Clustered Datastore [Feature:vsphere] verify dynamic provision with default parameter on clustered datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Verify Volume Attach Through vpxd Restart [Feature:vsphere][Serial][Disruptive] verify volume remains attached through vpxd restart" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Generic Ephemeral-volume (default fs) [Feature:GenericEphemeralVolume] (late-binding)] ephemeral should support multiple inline ephemeral volumes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes [Feature:vsphere][Feature:ReclaimPolicy] [sig-storage] persistentvolumereclaim:vsphere [Feature:vsphere] should retain persistent volume when reclaimPolicy set to retain when associated claim is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl create quota should reject quota with invalid scopes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Docker Containers should be able to override the image&#39;s default arguments (docker cmd) [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="4.069124729"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Probing container should *not* be restarted with a non-local redirect http liveness probe" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Container Runtime blackbox test when running a container with a new image should not be able to pull from private registry without secret [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory &#39;adir&#39; when HostPathType is HostPathBlockDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify PVC creation with compatible policy and datastore without any zones specified in the storage class fails (No shared datastores exist among all the nodes)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: gce-localssd-scsi-fs] [Serial] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI workload information using mock driver contain ephemeral=true when using inline volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should allow ingress access from updated pod [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStackAlphaFeature] [LinuxOnly] should create service with ipv4 cluster ip [Feature:IPv6DualStackAlphaFeature:Phase2]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link-bindmounted] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]" classname="Kubernetes e2e suite" time="13.315791988"></testcase>
      <testcase name="[sig-apps] DaemonRestart [Disruptive] Scheduler should continue assigning pods to nodes across restart" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : secret" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning DynamicProvisioner Default should be disabled by removing the default annotation [Serial] [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes NFS with multiple PVs and PVCs all in same ns should create 4 PVs and 2 PVCs: test write access [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] One pod requesting one prebound PVC should be able to mount volume and read from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Deployment should not disrupt a cloud load-balancer&#39;s connectivity during rollout" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should work after restarting apiserver [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Cluster level logging implemented by Stackdriver should ingest system logs from all nodes [Feature:StackdriverLogging]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Provisioning On Clustered Datastore [Feature:vsphere] verify dynamic provision with spbm policy on clustered datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStackAlphaFeature] [LinuxOnly] should create service with cluster ip from primary service range [Feature:IPv6DualStackAlphaFeature:Phase2]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Variable Expansion should allow substituting values in a container&#39;s args [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="6.065343612"></testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS should provide DNS for services  [Conformance]" classname="Kubernetes e2e suite" time="15.189231241"></testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance]" classname="Kubernetes e2e suite" time="0.083766855"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] Multi-AZ Cluster Volumes [sig-storage] should schedule pods in the same zones as statically provisioned PVs" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should scale up correct target pool [Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should allow ingress access from updated namespace [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: tmpfs] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Pod garbage collector [Feature:PodGarbageCollector] [Slow] should handle the creation of 1000 pods" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) [Feature:GenericEphemeralVolume] (immediate-binding)] ephemeral should support two pods which share the same volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should provide podname as non-root with fsgroup [LinuxOnly] [NodeFeature:FSGroup]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volumes ConfigMap should be mountable" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PV Protection Verify that PV bound to a PVC is not removed immediately" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Placement [Feature:vsphere] should create and delete pod with multiple volumes from same datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] [Feature:NodeAuthorizer] A node shouldn&#39;t be able to create another node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  StatefulSet with pod affinity [Slow] should use volumes on one node when pod management is parallel and pod has affinity" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV, based on the allowed zones specified in storage class when the datastore under the zone is present in another datacenter" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Cluster size autoscaler scalability [Slow] should scale down underutilized nodes [Feature:ClusterAutoscalerScalability4]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] [Feature:NodeAuthenticator] The kubelet can delegate ServiceAccount tokens to the API server" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume FStype [Feature:vsphere] verify invalid fstype" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]" classname="Kubernetes e2e suite" time="25.079284088"></testcase>
      <testcase name="[sig-storage] PersistentVolumes [Feature:vsphere][Feature:LabelSelector] [sig-storage] Selector-Label Volume Binding:vsphere [Feature:vsphere] should bind volume with claim for given label" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Probing container should be restarted with a local redirect http liveness probe" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]" classname="Kubernetes e2e suite" time="15.621815805"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStackAlphaFeature] [LinuxOnly] should be able to reach pod on ipv4 and ipv6 ip [Feature:IPv6DualStackAlphaFeature:Phase2]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] [Feature:NodeAuthorizer] Getting a non-existent secret should exit with the Forbidden error, not a NotFound error" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device &#39;ablkdev&#39; when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning DynamicProvisioner Default should be disabled by changing the default annotation [Serial] [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] MetricsGrabber should grab all metrics from API server." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume storage capacity exhausted, late binding, no topology" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="150.296448254"></testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates that taints-tolerations is respected if not matching" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS configMap nameserver Forward PTR lookup should forward PTR records lookup to upstream nameserver [Slow][Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable[Feature:VolumeSnapshotDataSource] volume snapshot controller  should check snapshot fields, check restore correctly works after modifying source data, check deletion" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be able to update service type to NodePort listening on same port number but different protocols" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:kubemci] single and multi-cluster ingresses should be able to exist together" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] DNS should support configurable pod DNS servers" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="6.057298126"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl get componentstatuses should get componentstatuses" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS should support configurable pod resolv.conf" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify PVC creation with incompatible datastore and zone combination specified in storage class fails" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Pods should have their auto-restart back-off timer reset on image update [Slow][NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should correctly scale down after a node is not needed when there is non autoscaled pool[Feature:ClusterSizeAutoscalingScaleDown]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Stackdriver Metadata Agent [Feature:StackdriverMetadataAgent]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Lease lease API should be available [Conformance]" classname="Kubernetes e2e suite" time="0.075584072"></testcase>
      <testcase name="[k8s.io] [sig-node] Security Context should support seccomp default which is unconfined [Feature:Seccomp] [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]" classname="Kubernetes e2e suite" time="5.592394445"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] Set fsGroup for local volume should set fsGroup for one pod [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="51.082054362"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="29.990502202000002"></testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Socket [Slow] Should fail on mounting socket &#39;asocket&#39; when HostPathType is HostPathDirectory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] files with FSGroup ownership should support (root,0644,tmpfs)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes GCEPD should test that deleting the PV before the pod does not cause pod deletion to fail on PD detach" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Custom Metrics - Stackdriver Adapter for old resource model [Feature:StackdriverCustomMetrics]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] [Serial] Volume metrics should create metrics for total time taken in volume operations in P/V Controller" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking IPerf IPv4 [Experimental] [Feature:Networking-IPv4] [Slow] [Feature:Networking-Performance] should transfer ~ 1GB onto the service endpoint 1 servers (maximum of 1 clients)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Socket [Slow] Should be able to mount socket &#39;asocket&#39; successfully when HostPathType is HostPathUnset" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="914.12692366">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597&#xA;May  4 16:25:53.905: Container should have service environment variables set&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0028b8490&gt;: {&#xA;        s: &#34;expected pod \&#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49\&#34; success: Gave up after waiting 5m0s for pod \&#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49\&#34; to be \&#34;Succeeded or Failed\&#34;&#34;,&#xA;    }&#xA;    expected pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; success: Gave up after waiting 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; to be &#34;Succeeded or Failed&#34;&#xA;occurred&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:527</failure>
          <system-out>[BeforeEach] [k8s.io] Pods&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174&#xA;STEP: Creating a kubernetes client&#xA;May  4 16:10:40.128: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename pods&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[BeforeEach] [k8s.io] Pods&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181&#xA;[It] should contain environment variables for services [NodeConformance] [Conformance]&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597&#xA;May  4 16:10:44.190: INFO: Waiting up to 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; in namespace &#34;pods-2302&#34; to be &#34;Succeeded or Failed&#34;&#xA;May  4 16:10:44.192: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2.387737ms&#xA;May  4 16:10:46.195: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2.005048639s&#xA;May  4 16:10:48.198: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4.008330081s&#xA;May  4 16:10:50.203: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 6.01292739s&#xA;May  4 16:10:52.207: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 8.01688963s&#xA;May  4 16:10:54.211: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 10.021554297s&#xA;May  4 16:10:56.215: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 12.024915137s&#xA;May  4 16:10:58.218: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 14.027970856s&#xA;May  4 16:11:00.221: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 16.031271402s&#xA;May  4 16:11:02.224: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 18.034560075s&#xA;May  4 16:11:04.230: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 20.040597735s&#xA;May  4 16:11:06.233: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 22.043630728s&#xA;May  4 16:11:08.237: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 24.047617964s&#xA;May  4 16:11:10.241: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 26.050933979s&#xA;May  4 16:11:12.246: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 28.055839793s&#xA;May  4 16:11:14.250: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 30.059768635s&#xA;May  4 16:11:16.253: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 32.062840352s&#xA;May  4 16:11:18.257: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 34.067105498s&#xA;May  4 16:11:20.260: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 36.070155248s&#xA;May  4 16:11:22.263: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 38.072693521s&#xA;May  4 16:11:24.265: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 40.075599917s&#xA;May  4 16:11:26.269: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 42.07914466s&#xA;May  4 16:11:28.273: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 44.082830404s&#xA;May  4 16:11:30.277: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 46.086680807s&#xA;May  4 16:11:32.279: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 48.089122005s&#xA;May  4 16:11:34.283: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 50.09270859s&#xA;May  4 16:11:36.286: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 52.096635192s&#xA;May  4 16:11:38.289: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 54.099058174s&#xA;May  4 16:11:40.292: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 56.102372708s&#xA;May  4 16:11:42.295: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 58.104707357s&#xA;May  4 16:11:44.298: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m0.108164815s&#xA;May  4 16:11:46.301: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m2.111079673s&#xA;May  4 16:11:48.304: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m4.114158777s&#xA;May  4 16:11:50.307: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m6.117332077s&#xA;May  4 16:11:52.310: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m8.120166469s&#xA;May  4 16:11:54.314: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m10.12425816s&#xA;May  4 16:11:56.318: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m12.12765473s&#xA;May  4 16:11:58.322: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m14.132620748s&#xA;May  4 16:12:00.326: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m16.136250965s&#xA;May  4 16:12:02.329: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m18.139538752s&#xA;May  4 16:12:04.336: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m20.145906521s&#xA;May  4 16:12:06.340: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m22.150502445s&#xA;May  4 16:12:08.344: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m24.154579048s&#xA;May  4 16:12:10.348: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m26.15859473s&#xA;May  4 16:12:12.351: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m28.161278768s&#xA;May  4 16:12:14.355: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m30.165502642s&#xA;May  4 16:12:16.359: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m32.169294447s&#xA;May  4 16:12:18.362: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m34.172582128s&#xA;May  4 16:12:20.367: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m36.177632595s&#xA;May  4 16:12:22.371: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m38.181363272s&#xA;May  4 16:12:24.375: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m40.18502512s&#xA;May  4 16:12:26.378: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m42.188164789s&#xA;May  4 16:12:28.382: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m44.191772996s&#xA;May  4 16:12:30.387: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m46.196949124s&#xA;May  4 16:12:32.390: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m48.200486048s&#xA;May  4 16:12:34.394: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m50.204426236s&#xA;May  4 16:12:36.399: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m52.208931419s&#xA;May  4 16:12:38.404: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m54.213679041s&#xA;May  4 16:12:40.409: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m56.219369851s&#xA;May  4 16:12:42.412: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m58.222533904s&#xA;May  4 16:12:44.415: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m0.22530163s&#xA;May  4 16:12:46.418: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m2.228198071s&#xA;May  4 16:12:48.422: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m4.231944139s&#xA;May  4 16:12:50.424: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m6.234530682s&#xA;May  4 16:12:52.427: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m8.237340984s&#xA;May  4 16:12:54.430: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m10.240104216s&#xA;May  4 16:12:56.433: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m12.243209422s&#xA;May  4 16:12:58.436: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m14.245772999s&#xA;May  4 16:13:00.439: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m16.249389394s&#xA;May  4 16:13:02.442: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m18.252588877s&#xA;May  4 16:13:04.446: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m20.256164321s&#xA;May  4 16:13:06.449: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m22.258966285s&#xA;May  4 16:13:08.453: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m24.262705925s&#xA;May  4 16:13:10.457: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m26.267198143s&#xA;May  4 16:13:12.460: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m28.270464458s&#xA;May  4 16:13:14.464: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m30.274102883s&#xA;May  4 16:13:16.467: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m32.277574814s&#xA;May  4 16:13:18.471: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m34.281544017s&#xA;May  4 16:13:20.475: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m36.284757306s&#xA;May  4 16:13:22.478: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m38.287724425s&#xA;May  4 16:13:24.480: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m40.2903342s&#xA;May  4 16:13:26.483: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m42.293630885s&#xA;May  4 16:13:28.487: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m44.296702802s&#xA;May  4 16:13:30.490: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m46.30052558s&#xA;May  4 16:13:32.494: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m48.303966548s&#xA;May  4 16:13:34.498: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m50.307719687s&#xA;May  4 16:13:36.501: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m52.311542052s&#xA;May  4 16:13:38.504: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m54.314521164s&#xA;May  4 16:13:40.507: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m56.317596121s&#xA;May  4 16:13:42.510: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m58.320254572s&#xA;May  4 16:13:44.513: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m0.32292295s&#xA;May  4 16:13:46.515: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m2.325405239s&#xA;May  4 16:13:48.518: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m4.327938597s&#xA;May  4 16:13:50.522: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m6.332445539s&#xA;May  4 16:13:52.526: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m8.336638233s&#xA;May  4 16:13:54.530: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m10.340038397s&#xA;May  4 16:13:56.534: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m12.343953682s&#xA;May  4 16:13:58.538: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m14.348589509s&#xA;May  4 16:14:00.543: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m16.352894856s&#xA;May  4 16:14:02.547: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m18.357359982s&#xA;May  4 16:14:04.552: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m20.36232504s&#xA;May  4 16:14:06.556: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m22.365722217s&#xA;May  4 16:14:08.561: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m24.371027721s&#xA;May  4 16:14:10.566: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m26.375922743s&#xA;May  4 16:14:12.569: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m28.37915381s&#xA;May  4 16:14:14.573: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m30.38268475s&#xA;May  4 16:14:16.577: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m32.386722486s&#xA;May  4 16:14:18.582: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m34.392138653s&#xA;May  4 16:14:20.586: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m36.396499412s&#xA;May  4 16:14:22.589: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m38.399024185s&#xA;May  4 16:14:24.592: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m40.401975335s&#xA;May  4 16:14:26.597: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m42.407583303s&#xA;May  4 16:14:28.602: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m44.412315431s&#xA;May  4 16:14:30.607: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m46.416835334s&#xA;May  4 16:14:32.610: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m48.420140068s&#xA;May  4 16:14:34.613: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m50.423573141s&#xA;May  4 16:14:36.617: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m52.42732293s&#xA;May  4 16:14:38.621: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m54.43140031s&#xA;May  4 16:14:40.625: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m56.435439817s&#xA;May  4 16:14:42.628: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m58.438371798s&#xA;May  4 16:14:44.632: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m0.441810606s&#xA;May  4 16:14:46.636: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m2.446286891s&#xA;May  4 16:14:48.640: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m4.450469377s&#xA;May  4 16:14:50.644: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m6.454210918s&#xA;May  4 16:14:52.648: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m8.457954378s&#xA;May  4 16:14:54.651: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m10.460816818s&#xA;May  4 16:14:56.655: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m12.465070873s&#xA;May  4 16:14:58.658: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m14.468267874s&#xA;May  4 16:15:00.663: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m16.472908465s&#xA;May  4 16:15:02.667: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m18.477504538s&#xA;May  4 16:15:04.671: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m20.481066686s&#xA;May  4 16:15:06.674: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m22.483733149s&#xA;May  4 16:15:08.678: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m24.488496996s&#xA;May  4 16:15:10.683: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m26.492844859s&#xA;May  4 16:15:12.687: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m28.497194266s&#xA;May  4 16:15:14.691: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m30.500834246s&#xA;May  4 16:15:16.695: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m32.504800727s&#xA;May  4 16:15:18.698: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m34.508214412s&#xA;May  4 16:15:20.703: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m36.513036802s&#xA;May  4 16:15:22.708: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m38.518360261s&#xA;May  4 16:15:24.711: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m40.521553626s&#xA;May  4 16:15:26.715: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m42.525224056s&#xA;May  4 16:15:28.717: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m44.527625437s&#xA;May  4 16:15:30.720: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m46.530394579s&#xA;May  4 16:15:32.723: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m48.53362351s&#xA;May  4 16:15:34.727: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m50.536867334s&#xA;May  4 16:15:36.729: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m52.539616993s&#xA;May  4 16:15:38.734: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m54.543730357s&#xA;May  4 16:15:40.737: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m56.547375917s&#xA;May  4 16:15:42.742: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m58.552555981s&#xA;May  4 16:15:44.750: INFO: Failed to get logs from node &#34;node1&#34; pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; container &#34;env3cont&#34;: the server rejected our request for an unknown reason (get pods client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49)&#xA;STEP: delete the pod&#xA;May  4 16:15:44.754: INFO: Waiting for pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to disappear&#xA;May  4 16:15:44.757: INFO: Pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 still exists&#xA;May  4 16:15:46.757: INFO: Waiting for pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to disappear&#xA;May  4 16:15:46.760: INFO: Pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 no longer exists&#xA;May  4 16:15:46.760: INFO: (Attempt 1 of 3) Unexpected error occurred: expected pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; success: Gave up after waiting 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; to be &#34;Succeeded or Failed&#34;&#xA;May  4 16:15:46.776: INFO: Waiting up to 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; in namespace &#34;pods-2302&#34; to be &#34;Succeeded or Failed&#34;&#xA;May  4 16:15:46.778: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1.835657ms&#xA;May  4 16:15:48.781: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2.005026733s&#xA;May  4 16:15:50.785: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4.008454957s&#xA;May  4 16:15:52.788: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 6.011820889s&#xA;May  4 16:15:54.791: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 8.014849549s&#xA;May  4 16:15:56.796: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 10.019473952s&#xA;May  4 16:15:58.799: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 12.023061432s&#xA;May  4 16:16:00.804: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 14.027557964s&#xA;May  4 16:16:02.808: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 16.031860161s&#xA;May  4 16:16:04.811: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 18.035337128s&#xA;May  4 16:16:06.815: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 20.039167428s&#xA;May  4 16:16:08.818: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 22.042323917s&#xA;May  4 16:16:10.822: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 24.045630093s&#xA;May  4 16:16:12.826: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 26.049410161s&#xA;May  4 16:16:14.828: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 28.052112349s&#xA;May  4 16:16:16.832: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 30.055464809s&#xA;May  4 16:16:18.836: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 32.059704326s&#xA;May  4 16:16:20.839: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 34.062856973s&#xA;May  4 16:16:22.842: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 36.066274293s&#xA;May  4 16:16:24.846: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 38.069810421s&#xA;May  4 16:16:26.849: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 40.073345881s&#xA;May  4 16:16:28.854: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 42.077408007s&#xA;May  4 16:16:30.857: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 44.080744669s&#xA;May  4 16:16:32.860: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 46.084308543s&#xA;May  4 16:16:34.865: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 48.088786704s&#xA;May  4 16:16:36.869: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 50.092636957s&#xA;May  4 16:16:38.873: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 52.097150136s&#xA;May  4 16:16:40.876: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 54.09966338s&#xA;May  4 16:16:42.879: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 56.102480627s&#xA;May  4 16:16:44.881: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 58.10526284s&#xA;May  4 16:16:46.886: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m0.109958239s&#xA;May  4 16:16:48.890: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m2.113518074s&#xA;May  4 16:16:50.893: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m4.116940863s&#xA;May  4 16:16:52.897: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m6.120374958s&#xA;May  4 16:16:54.900: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m8.123404884s&#xA;May  4 16:16:56.902: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m10.126191732s&#xA;May  4 16:16:58.905: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m12.128717411s&#xA;May  4 16:17:00.907: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m14.131160185s&#xA;May  4 16:17:02.911: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m16.134829878s&#xA;May  4 16:17:04.915: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m18.138894601s&#xA;May  4 16:17:06.918: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m20.142202847s&#xA;May  4 16:17:08.922: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m22.145770341s&#xA;May  4 16:17:10.926: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m24.14970037s&#xA;May  4 16:17:12.930: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m26.153827539s&#xA;May  4 16:17:14.935: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m28.15861431s&#xA;May  4 16:17:16.938: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m30.161932902s&#xA;May  4 16:17:18.942: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m32.166239738s&#xA;May  4 16:17:20.946: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m34.169381617s&#xA;May  4 16:17:22.949: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m36.173053531s&#xA;May  4 16:17:24.955: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m38.178506724s&#xA;May  4 16:17:26.958: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m40.182066155s&#xA;May  4 16:17:28.962: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m42.186141678s&#xA;May  4 16:17:30.966: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m44.189436731s&#xA;May  4 16:17:32.970: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m46.193636904s&#xA;May  4 16:17:34.975: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m48.198464865s&#xA;May  4 16:17:36.978: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m50.201762981s&#xA;May  4 16:17:38.982: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m52.205763674s&#xA;May  4 16:17:40.985: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m54.208575129s&#xA;May  4 16:17:42.988: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m56.212121117s&#xA;May  4 16:17:44.993: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m58.216627023s&#xA;May  4 16:17:46.996: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m0.219931491s&#xA;May  4 16:17:49.000: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m2.223647711s&#xA;May  4 16:17:51.004: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m4.227476047s&#xA;May  4 16:17:53.008: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m6.232292879s&#xA;May  4 16:17:55.012: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m8.236094584s&#xA;May  4 16:17:57.018: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m10.241923389s&#xA;May  4 16:17:59.022: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m12.246247959s&#xA;May  4 16:18:01.025: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m14.249146435s&#xA;May  4 16:18:03.030: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m16.254293974s&#xA;May  4 16:18:05.039: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m18.262491524s&#xA;May  4 16:18:07.043: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m20.266773286s&#xA;May  4 16:18:09.048: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m22.271664138s&#xA;May  4 16:18:11.052: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m24.276130928s&#xA;May  4 16:18:13.055: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m26.279061412s&#xA;May  4 16:18:15.058: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m28.282339094s&#xA;May  4 16:18:17.062: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m30.285601705s&#xA;May  4 16:18:19.065: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m32.289113197s&#xA;May  4 16:18:21.068: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m34.29233631s&#xA;May  4 16:18:23.072: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m36.296100434s&#xA;May  4 16:18:25.075: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m38.299022812s&#xA;May  4 16:18:27.079: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m40.303130001s&#xA;May  4 16:18:29.082: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m42.30631831s&#xA;May  4 16:18:31.087: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m44.310709419s&#xA;May  4 16:18:33.091: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m46.314456553s&#xA;May  4 16:18:35.094: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m48.318108398s&#xA;May  4 16:18:37.097: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m50.320937618s&#xA;May  4 16:18:39.100: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m52.32404526s&#xA;May  4 16:18:41.103: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m54.326640659s&#xA;May  4 16:18:43.106: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m56.329637348s&#xA;May  4 16:18:45.110: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m58.333577382s&#xA;May  4 16:18:47.114: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m0.337647345s&#xA;May  4 16:18:49.117: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m2.341316532s&#xA;May  4 16:18:51.121: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m4.344744729s&#xA;May  4 16:18:53.124: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m6.34739326s&#xA;May  4 16:18:55.126: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m8.350238621s&#xA;May  4 16:18:57.131: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m10.355247669s&#xA;May  4 16:18:59.135: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m12.359090125s&#xA;May  4 16:19:01.139: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m14.363033458s&#xA;May  4 16:19:03.143: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m16.367035304s&#xA;May  4 16:19:05.148: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m18.372312379s&#xA;May  4 16:19:07.153: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m20.377362627s&#xA;May  4 16:19:09.156: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m22.379695785s&#xA;May  4 16:19:11.159: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m24.382773254s&#xA;May  4 16:19:13.162: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m26.386316986s&#xA;May  4 16:19:15.166: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m28.389778748s&#xA;May  4 16:19:17.169: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m30.39247196s&#xA;May  4 16:19:19.172: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m32.396337189s&#xA;May  4 16:19:21.175: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m34.399006966s&#xA;May  4 16:19:23.179: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m36.402445564s&#xA;May  4 16:19:25.182: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m38.405460729s&#xA;May  4 16:19:27.185: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m40.409102635s&#xA;May  4 16:19:29.188: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m42.411565405s&#xA;May  4 16:19:31.192: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m44.415755748s&#xA;May  4 16:19:33.195: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m46.418373695s&#xA;May  4 16:19:35.198: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m48.421603651s&#xA;May  4 16:19:37.202: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m50.425641397s&#xA;May  4 16:19:39.204: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m52.428359815s&#xA;May  4 16:19:41.208: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m54.431531578s&#xA;May  4 16:19:43.211: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m56.434976551s&#xA;May  4 16:19:45.215: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m58.439204503s&#xA;May  4 16:19:47.218: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m0.441578552s&#xA;May  4 16:19:49.225: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m2.449060294s&#xA;May  4 16:19:51.229: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m4.452758171s&#xA;May  4 16:19:53.232: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m6.456370017s&#xA;May  4 16:19:55.239: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m8.462538761s&#xA;May  4 16:19:57.245: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m10.469040256s&#xA;May  4 16:19:59.249: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m12.472415135s&#xA;May  4 16:20:01.252: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m14.475466656s&#xA;May  4 16:20:03.255: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m16.479320127s&#xA;May  4 16:20:05.260: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m18.483693837s&#xA;May  4 16:20:07.263: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m20.487231771s&#xA;May  4 16:20:09.266: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m22.490019564s&#xA;May  4 16:20:11.269: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m24.492796351s&#xA;May  4 16:20:13.271: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m26.495145335s&#xA;May  4 16:20:15.275: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m28.498524047s&#xA;May  4 16:20:17.278: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m30.501573201s&#xA;May  4 16:20:19.282: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m32.505594181s&#xA;May  4 16:20:21.284: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m34.508352419s&#xA;May  4 16:20:23.287: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m36.511283496s&#xA;May  4 16:20:25.291: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m38.514698773s&#xA;May  4 16:20:27.293: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m40.517289094s&#xA;May  4 16:20:29.296: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m42.520221689s&#xA;May  4 16:20:31.300: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m44.523449923s&#xA;May  4 16:20:33.302: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m46.526261268s&#xA;May  4 16:20:35.305: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m48.529262831s&#xA;May  4 16:20:37.309: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m50.532963329s&#xA;May  4 16:20:39.313: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m52.536438042s&#xA;May  4 16:20:41.316: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m54.539568908s&#xA;May  4 16:20:43.319: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m56.543022667s&#xA;May  4 16:20:45.323: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m58.546617957s&#xA;May  4 16:20:47.332: INFO: Failed to get logs from node &#34;node2&#34; pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; container &#34;env3cont&#34;: the server rejected our request for an unknown reason (get pods client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49)&#xA;STEP: delete the pod&#xA;May  4 16:20:47.339: INFO: Waiting for pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to disappear&#xA;May  4 16:20:47.341: INFO: Pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 still exists&#xA;May  4 16:20:49.342: INFO: Waiting for pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to disappear&#xA;May  4 16:20:49.344: INFO: Pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 no longer exists&#xA;May  4 16:20:49.344: INFO: (Attempt 2 of 3) Unexpected error occurred: expected pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; success: Gave up after waiting 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; to be &#34;Succeeded or Failed&#34;&#xA;May  4 16:20:49.356: INFO: Waiting up to 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; in namespace &#34;pods-2302&#34; to be &#34;Succeeded or Failed&#34;&#xA;May  4 16:20:49.358: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1.796197ms&#xA;May  4 16:20:51.362: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2.005555726s&#xA;May  4 16:20:53.366: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4.009742011s&#xA;May  4 16:20:55.369: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 6.012336288s&#xA;May  4 16:20:57.373: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 8.016026391s&#xA;May  4 16:20:59.376: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 10.019780226s&#xA;May  4 16:21:01.379: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 12.022664708s&#xA;May  4 16:21:03.383: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 14.026172542s&#xA;May  4 16:21:05.387: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 16.030475438s&#xA;May  4 16:21:07.391: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 18.034522314s&#xA;May  4 16:21:09.394: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 20.037862353s&#xA;May  4 16:21:11.398: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 22.041444928s&#xA;May  4 16:21:13.402: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 24.045225223s&#xA;May  4 16:21:15.404: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 26.047807644s&#xA;May  4 16:21:17.407: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 28.050976487s&#xA;May  4 16:21:19.410: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 30.053227357s&#xA;May  4 16:21:21.412: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 32.055737601s&#xA;May  4 16:21:23.417: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 34.060584834s&#xA;May  4 16:21:25.420: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 36.063818944s&#xA;May  4 16:21:27.424: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 38.067731095s&#xA;May  4 16:21:29.428: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 40.071598153s&#xA;May  4 16:21:31.432: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 42.075034612s&#xA;May  4 16:21:33.435: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 44.078246096s&#xA;May  4 16:21:35.438: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 46.081608055s&#xA;May  4 16:21:37.441: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 48.084681733s&#xA;May  4 16:21:39.445: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 50.088150488s&#xA;May  4 16:21:41.447: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 52.090931058s&#xA;May  4 16:21:43.450: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 54.093964716s&#xA;May  4 16:21:45.453: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 56.096831449s&#xA;May  4 16:21:47.457: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 58.100298272s&#xA;May  4 16:21:49.460: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m0.103630661s&#xA;May  4 16:21:51.465: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m2.108547157s&#xA;May  4 16:21:53.470: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m4.113647337s&#xA;May  4 16:21:55.475: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m6.118962603s&#xA;May  4 16:21:57.479: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m8.12284588s&#xA;May  4 16:21:59.484: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m10.127640541s&#xA;May  4 16:22:01.487: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m12.130292308s&#xA;May  4 16:22:03.492: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m14.135845336s&#xA;May  4 16:22:05.496: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m16.139739371s&#xA;May  4 16:22:07.500: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m18.143930979s&#xA;May  4 16:22:09.505: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m20.1481924s&#xA;May  4 16:22:11.508: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m22.151890847s&#xA;May  4 16:22:13.513: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m24.156516911s&#xA;May  4 16:22:15.516: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m26.159406197s&#xA;May  4 16:22:17.518: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m28.161902329s&#xA;May  4 16:22:19.521: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m30.164613992s&#xA;May  4 16:22:21.525: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m32.168246673s&#xA;May  4 16:22:23.529: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m34.172602702s&#xA;May  4 16:22:25.533: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m36.176266304s&#xA;May  4 16:22:27.536: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m38.179241045s&#xA;May  4 16:22:29.539: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m40.18286735s&#xA;May  4 16:22:31.543: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m42.186987483s&#xA;May  4 16:22:33.547: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m44.1902214s&#xA;May  4 16:22:35.550: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m46.193386449s&#xA;May  4 16:22:37.553: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m48.19639727s&#xA;May  4 16:22:39.556: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m50.199351157s&#xA;May  4 16:22:41.559: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m52.202403336s&#xA;May  4 16:22:43.562: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m54.205406523s&#xA;May  4 16:22:45.565: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m56.208082268s&#xA;May  4 16:22:47.568: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 1m58.211188938s&#xA;May  4 16:22:49.571: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m0.214432801s&#xA;May  4 16:22:51.575: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m2.218139448s&#xA;May  4 16:22:53.579: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m4.222333048s&#xA;May  4 16:22:55.582: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m6.225497302s&#xA;May  4 16:22:57.586: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m8.229603116s&#xA;May  4 16:22:59.589: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m10.232921493s&#xA;May  4 16:23:01.593: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m12.236750713s&#xA;May  4 16:23:03.596: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m14.239742147s&#xA;May  4 16:23:05.600: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m16.243239875s&#xA;May  4 16:23:07.603: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m18.246494059s&#xA;May  4 16:23:09.606: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m20.249879762s&#xA;May  4 16:23:11.610: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m22.253873574s&#xA;May  4 16:23:13.613: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m24.256866255s&#xA;May  4 16:23:15.616: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m26.25993074s&#xA;May  4 16:23:17.620: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m28.26333395s&#xA;May  4 16:23:19.623: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m30.26689884s&#xA;May  4 16:23:21.626: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m32.269724618s&#xA;May  4 16:23:23.630: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m34.273442909s&#xA;May  4 16:23:25.634: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m36.277881534s&#xA;May  4 16:23:27.637: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m38.280589047s&#xA;May  4 16:23:29.640: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m40.283215785s&#xA;May  4 16:23:31.643: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m42.286019766s&#xA;May  4 16:23:33.645: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m44.288838124s&#xA;May  4 16:23:35.649: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m46.292436002s&#xA;May  4 16:23:37.653: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m48.296007066s&#xA;May  4 16:23:39.655: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m50.298896131s&#xA;May  4 16:23:41.659: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m52.30226912s&#xA;May  4 16:23:43.663: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m54.306223975s&#xA;May  4 16:23:45.666: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m56.309127024s&#xA;May  4 16:23:47.670: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 2m58.313409924s&#xA;May  4 16:23:49.673: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m0.316735968s&#xA;May  4 16:23:51.676: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m2.319201997s&#xA;May  4 16:23:53.679: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m4.322790157s&#xA;May  4 16:23:55.684: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m6.327174201s&#xA;May  4 16:23:57.688: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m8.331024421s&#xA;May  4 16:23:59.691: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m10.334802798s&#xA;May  4 16:24:01.694: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m12.33713604s&#xA;May  4 16:24:03.697: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m14.340701704s&#xA;May  4 16:24:05.700: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m16.343522975s&#xA;May  4 16:24:07.703: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m18.346965214s&#xA;May  4 16:24:09.706: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m20.34936454s&#xA;May  4 16:24:11.709: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m22.352128106s&#xA;May  4 16:24:13.711: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m24.354941099s&#xA;May  4 16:24:15.715: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m26.358236872s&#xA;May  4 16:24:17.718: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m28.361715158s&#xA;May  4 16:24:19.723: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m30.36609709s&#xA;May  4 16:24:21.725: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m32.369001164s&#xA;May  4 16:24:23.730: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m34.373840001s&#xA;May  4 16:24:25.734: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m36.37711331s&#xA;May  4 16:24:27.736: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m38.379558979s&#xA;May  4 16:24:29.739: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m40.382964113s&#xA;May  4 16:24:31.742: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m42.385935629s&#xA;May  4 16:24:33.747: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m44.390041799s&#xA;May  4 16:24:35.751: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m46.394115706s&#xA;May  4 16:24:37.754: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m48.397123922s&#xA;May  4 16:24:39.757: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m50.400844108s&#xA;May  4 16:24:41.761: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m52.404674162s&#xA;May  4 16:24:43.765: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m54.408907042s&#xA;May  4 16:24:45.770: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m56.413028656s&#xA;May  4 16:24:47.773: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 3m58.41619143s&#xA;May  4 16:24:49.777: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m0.420908528s&#xA;May  4 16:24:51.782: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m2.425285433s&#xA;May  4 16:24:53.786: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m4.429325996s&#xA;May  4 16:24:55.789: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m6.432155219s&#xA;May  4 16:24:57.793: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m8.436309758s&#xA;May  4 16:24:59.796: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m10.439791238s&#xA;May  4 16:25:01.799: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m12.442412965s&#xA;May  4 16:25:03.803: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m14.44687452s&#xA;May  4 16:25:05.807: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m16.450018833s&#xA;May  4 16:25:07.811: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m18.454068829s&#xA;May  4 16:25:09.814: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m20.457361954s&#xA;May  4 16:25:11.817: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m22.460875554s&#xA;May  4 16:25:13.821: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m24.46469586s&#xA;May  4 16:25:15.824: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m26.467997832s&#xA;May  4 16:25:17.827: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m28.470469957s&#xA;May  4 16:25:19.831: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m30.474166645s&#xA;May  4 16:25:21.834: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m32.477942873s&#xA;May  4 16:25:23.837: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m34.480790506s&#xA;May  4 16:25:25.841: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m36.4845894s&#xA;May  4 16:25:27.845: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m38.488018609s&#xA;May  4 16:25:29.848: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m40.491318177s&#xA;May  4 16:25:31.851: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m42.494271653s&#xA;May  4 16:25:33.856: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m44.499246378s&#xA;May  4 16:25:35.860: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m46.503036032s&#xA;May  4 16:25:37.864: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m48.507444248s&#xA;May  4 16:25:39.867: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m50.510334286s&#xA;May  4 16:25:41.871: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m52.514935217s&#xA;May  4 16:25:43.876: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m54.51999705s&#xA;May  4 16:25:45.879: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m56.522948483s&#xA;May  4 16:25:47.882: INFO: Pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34;: Phase=&#34;Pending&#34;, Reason=&#34;&#34;, readiness=false. Elapsed: 4m58.525991384s&#xA;May  4 16:25:49.891: INFO: Failed to get logs from node &#34;node1&#34; pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; container &#34;env3cont&#34;: the server rejected our request for an unknown reason (get pods client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49)&#xA;STEP: delete the pod&#xA;May  4 16:25:49.897: INFO: Waiting for pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to disappear&#xA;May  4 16:25:49.899: INFO: Pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 still exists&#xA;May  4 16:25:51.902: INFO: Waiting for pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to disappear&#xA;May  4 16:25:51.907: INFO: Pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 still exists&#xA;May  4 16:25:53.901: INFO: Waiting for pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to disappear&#xA;May  4 16:25:53.904: INFO: Pod client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 no longer exists&#xA;May  4 16:25:53.904: INFO: (Attempt 3 of 3) Unexpected error occurred: expected pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; success: Gave up after waiting 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; to be &#34;Succeeded or Failed&#34;&#xA;goroutine 217 [running]:&#xA;runtime/debug.Stack(0x4, 0x4be5f6e, 0x2)&#xA;&#x9;/usr/local/go/src/runtime/debug/stack.go:24 +0x9f&#xA;runtime/debug.PrintStack()&#xA;&#x9;/usr/local/go/src/runtime/debug/stack.go:16 +0x25&#xA;k8s.io/kubernetes/test/e2e/common.expectNoErrorWithRetries(0xc001a67108, 0x3, 0xc004e11900, 0x1, 0x1)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:173 +0x2bc&#xA;k8s.io/kubernetes/test/e2e/common.glob..func18.6()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:527 +0xaf2&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/leafnodes.(*runner).runSync(0xc000c1ac00, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/leafnodes/runner.go:113 +0xa3&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/leafnodes.(*runner).run(0xc000c1ac00, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/leafnodes/runner.go:64 +0x15c&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/leafnodes.(*ItNode).Run(0xc0007bf400, 0x52e17e0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/leafnodes/it_node.go:26 +0x87&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/spec.(*Spec).runSample(0xc0040baff0, 0x0, 0x52e17e0, 0xc000190900)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/spec/spec.go:215 +0x72f&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/spec.(*Spec).Run(0xc0040baff0, 0x52e17e0, 0xc000190900)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/spec/spec.go:138 +0xf2&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).runSpec(0xc00169e000, 0xc0040baff0, 0x0)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:200 +0x111&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).runSpecs(0xc00169e000, 0x1)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:170 +0x127&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).Run(0xc00169e000, 0xc002f5c030)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:66 +0x117&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/suite.(*Suite).Run(0xc000194280, 0x7efd15c06970, 0xc001827080, 0x4c22012, 0x14, 0xc0039a57a0, 0x3, 0x3, 0x5396840, 0xc000190900, ...)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/suite/suite.go:62 +0x426&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo.RunSpecsWithCustomReporters(0x52e6440, 0xc001827080, 0x4c22012, 0x14, 0xc000eb0240, 0x3, 0x4, 0x4)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/ginkgo_dsl.go:226 +0x238&#xA;k8s.io/kubernetes/vendor/github.com/onsi/ginkgo.RunSpecsWithDefaultAndCustomReporters(0x52e6440, 0xc001827080, 0x4c22012, 0x14, 0xc002452ca0, 0x2, 0x2, 0x2)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/ginkgo_dsl.go:214 +0xad&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc001827080)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc001827080)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc001827080, 0x4de37a0)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;May  4 16:25:53.905: FAIL: Container should have service environment variables set&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0028b8490&gt;: {&#xA;        s: &#34;expected pod \&#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49\&#34; success: Gave up after waiting 5m0s for pod \&#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49\&#34; to be \&#34;Succeeded or Failed\&#34;&#34;,&#xA;    }&#xA;    expected pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; success: Gave up after waiting 5m0s for pod &#34;client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49&#34; to be &#34;Succeeded or Failed&#34;&#xA;occurred&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/common.glob..func18.6()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:527 +0xaf2&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc001827080)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc001827080)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc001827080, 0x4de37a0)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;[AfterEach] [k8s.io] Pods&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175&#xA;STEP: Collecting events from namespace &#34;pods-2302&#34;.&#xA;STEP: Found 31 events.&#xA;May  4 16:25:53.910: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: { } Scheduled: Successfully assigned pods-2302/client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to node2&#xA;May  4 16:25:53.910: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: { } Scheduled: Successfully assigned pods-2302/client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to node1&#xA;May  4 16:25:53.910: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: { } Scheduled: Successfully assigned pods-2302/client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49 to node1&#xA;May  4 16:25:53.910: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4: { } Scheduled: Successfully assigned pods-2302/server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4 to node1&#xA;May  4 16:25:53.910: INFO: At 2021-05-04 16:10:41 +0000 UTC - event for server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4: {kubelet node1} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34;&#xA;May  4 16:25:53.910: INFO: At 2021-05-04 16:10:41 +0000 UTC - event for server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4: {multus } AddedInterface: Add eth0 [10.244.4.143/24]&#xA;May  4 16:25:53.910: INFO: At 2021-05-04 16:10:42 +0000 UTC - event for server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4: {kubelet node1} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34; in 450.386435ms&#xA;May  4 16:25:53.910: INFO: At 2021-05-04 16:10:42 +0000 UTC - event for server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4: {kubelet node1} Created: Created container srv&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:42 +0000 UTC - event for server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4: {kubelet node1} Started: Started container srv&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:45 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:45 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {multus } AddedInterface: Add eth0 [10.244.4.144/24]&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:46 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Failed: Error: ErrImagePull&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:46 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:47 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:48 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:48 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:10:48 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {multus } AddedInterface: Add eth0 [10.244.4.145/24]&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:15:48 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:15:48 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {multus } AddedInterface: Add eth0 [10.244.3.205/24]&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:15:49 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:15:49 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:15:49 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node2} Failed: Error: ErrImagePull&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:15:49 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:50 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {multus } AddedInterface: Add eth0 [10.244.4.176/24]&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:50 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:51 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Failed: Error: ErrImagePull&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:51 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:52 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:54 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {multus } AddedInterface: Add eth0 [10.244.4.178/24]&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:54 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:25:53.911: INFO: At 2021-05-04 16:20:54 +0000 UTC - event for client-envvars-3aca01f8-0d22-4951-b441-fa131ddecb49: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May  4 16:25:53.913: INFO: POD                                                  NODE   PHASE    GRACE  CONDITIONS&#xA;May  4 16:25:53.913: INFO: server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4  node1  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:10:40 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:10:43 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:10:43 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:10:40 +0000 UTC  }]&#xA;May  4 16:25:53.913: INFO: &#xA;May  4 16:25:53.917: INFO: &#xA;Logging node info for node master1&#xA;May  4 16:25:53.920: INFO: Node Info: &amp;Node{ObjectMeta:{master1   /api/v1/nodes/master1 db982204-549e-4532-90a7-a4410878cfc9 42260 0 2021-05-04 14:43:01 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;3e:f0:43:cb:66:52&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-04 14:43:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:43:02 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-04 14:45:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {nfd-master Update v1 2021-05-04 14:51:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:46 +0000 UTC,LastTransitionTime:2021-05-04 14:47:46 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:01 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:01 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:01 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:25:50 +0000 UTC,LastTransitionTime:2021-05-04 14:47:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:88a0771919594d4187f6704fc7592bf8,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:8e0a253b-2aa4-4467-879e-567e7ba1ffa4,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:c2fabacbc4e42f3db70f9508e00158b1dce4cf96d91cabaa2eca24e5a0900b66 tas-controller:latest localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[nginx@sha256:a97eb9ecc708c8aa715ccfb5e9338f5456e4b65575daf304f108301f3b497314 nginx:1.19.2-alpine],SizeBytes:22052669,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:e074a505e2d62b5119460ab724b2e1df10c8419ef2457f9ce9f3a0f75be3e959 tas-extender:latest localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:25:53.920: INFO: &#xA;Logging kubelet events for node master1&#xA;May  4 16:25:53.922: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May  4 16:25:53.945: INFO: kube-multus-ds-amd64-jflvf started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:25:53.945: INFO: coredns-7677f9bb54-qvcd2 started at 2021-05-04 14:46:11 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May  4 16:25:53.945: INFO: node-feature-discovery-controller-5bf5c49849-72rn6 started at 2021-05-04 14:51:52 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: kube-flannel-qspzk started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: &#x9;Container kube-flannel ready: true, restart count 3&#xA;May  4 16:25:53.945: INFO: kube-proxy-8j6ch started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May  4 16:25:53.945: INFO: docker-registry-docker-registry-56cbc7bc58-zhf8t started at 2021-05-04 14:48:42 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: node-exporter-jckjs started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: kube-scheduler-master1 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: kube-apiserver-master1 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May  4 16:25:53.945: INFO: kube-controller-manager-master1 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.945: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;W0504 16:25:53.957382      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:25:53.986: INFO: &#xA;Latency metrics for node master1&#xA;May  4 16:25:53.986: INFO: &#xA;Logging node info for node master2&#xA;May  4 16:25:53.989: INFO: Node Info: &amp;Node{ObjectMeta:{master2   /api/v1/nodes/master2 e2c15170-247b-4e7b-b818-abc807948bf8 42252 0 2021-05-04 14:43:40 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;16:e0:10:a0:e0:62&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-04 14:43:40 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:43:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-04 14:45:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:32 +0000 UTC,LastTransitionTime:2021-05-04 14:47:32 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:43:40 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:43:40 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:43:40 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:47:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:6af568f56589422a9bd68e0270ce0f8c,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:bf27bb77-fad2-4b52-85c3-acb5113fc512,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:25:53.989: INFO: &#xA;Logging kubelet events for node master2&#xA;May  4 16:25:53.991: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May  4 16:25:53.999: INFO: kube-proxy-6b5t8 started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May  4 16:25:53.999: INFO: kube-flannel-cxdfr started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May  4 16:25:53.999: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May  4 16:25:53.999: INFO: kube-multus-ds-amd64-dw8tg started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:25:53.999: INFO: dns-autoscaler-5b7b5c9b6f-zbrsq started at 2021-05-04 14:46:08 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;May  4 16:25:53.999: INFO: node-exporter-9c6qf started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:53.999: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:25:53.999: INFO: kube-apiserver-master2 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May  4 16:25:53.999: INFO: kube-controller-manager-master2 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May  4 16:25:53.999: INFO: kube-scheduler-master2 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:53.999: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;W0504 16:25:54.011625      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:25:54.034: INFO: &#xA;Latency metrics for node master2&#xA;May  4 16:25:54.035: INFO: &#xA;Logging node info for node master3&#xA;May  4 16:25:54.037: INFO: Node Info: &amp;Node{ObjectMeta:{master3   /api/v1/nodes/master3 b533a646-667e-403c-944b-71dec9cc4851 42251 0 2021-05-04 14:43:51 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;16:b0:53:14:f6:c9&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-04 14:43:51 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:43:52 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-04 14:45:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:48:50 +0000 UTC,LastTransitionTime:2021-05-04 14:48:50 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:43:51 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:43:51 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:43:51 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:25:49 +0000 UTC,LastTransitionTime:2021-05-04 14:45:46 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:959373dcda56494486f0c2bb0bb496cc,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:0714ca81-c21e-40d6-a288-48d597238e54,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:25:54.037: INFO: &#xA;Logging kubelet events for node master3&#xA;May  4 16:25:54.039: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May  4 16:25:54.047: INFO: kube-proxy-2p5b6 started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May  4 16:25:54.047: INFO: kube-flannel-wznt8 started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May  4 16:25:54.047: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May  4 16:25:54.047: INFO: kube-multus-ds-amd64-cgwz2 started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:25:54.047: INFO: coredns-7677f9bb54-pshfb started at 2021-05-04 14:46:06 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May  4 16:25:54.047: INFO: node-exporter-wvppn started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:54.047: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:25:54.047: INFO: kube-apiserver-master3 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May  4 16:25:54.047: INFO: kube-controller-manager-master3 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May  4 16:25:54.047: INFO: kube-scheduler-master3 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.047: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;W0504 16:25:54.060367      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:25:54.087: INFO: &#xA;Latency metrics for node master3&#xA;May  4 16:25:54.087: INFO: &#xA;Logging node info for node node1&#xA;May  4 16:25:54.090: INFO: Node Info: &amp;Node{ObjectMeta:{node1   /api/v1/nodes/node1 b8cf7e16-d5c7-4e2c-996a-93d93bd4fa1c 42239 0 2021-05-04 14:44:58 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;e2:50:df:03:d2:13&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-04 14:44:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:44:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-04 14:52:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-04 14:54:35 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-04 14:54:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:29 +0000 UTC,LastTransitionTime:2021-05-04 14:47:29 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:45 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:45 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:45 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:25:45 +0000 UTC,LastTransitionTime:2021-05-04 14:47:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:bffc023a4ab84df0b0181bc7b8f509e2,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:dc08af72-abca-4f1d-bd0f-0e8d8eb97de5,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:511d36b62d758304a5adb26b5996ed37211ab226beb7de4e67994cbecb0279a7 localhost:30500/barometer-collectd:stable],SizeBytes:1464048999,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1002569035,},ContainerImage{Names:[opnfv/barometer-collectd@sha256:ed5c574f653e2a39e784ff322033a2319aafde7366c803a88f20f7a2a8bc1efb opnfv/barometer-collectd:stable],SizeBytes:825413035,},ContainerImage{Names:[localhost:30500/cmk@sha256:f417461c5e0283b5f2ba8e34dc073a15fe1f9ff6b542330c536c86aa72f7141f cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:371dc6bf7e0c7ce112a29341b000c40d840aef1dbb4fdcb3ae5c0597e28f3061 golang:alpine3.12],SizeBytes:301097267,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[grafana/grafana@sha256:89304bc2335f4976618548d7b93d165ed67369d3a051d2f627fc4e0aa3d0aff1 grafana/grafana:7.1.0],SizeBytes:179601493,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:d4ba4dd1a9ebb90916d0bfed3c204adcb118ed24546bf8dd2e6b30fc0fd2009e quay.io/prometheus/prometheus:v2.20.0],SizeBytes:144886595,},ContainerImage{Names:[nginx@sha256:75a55d33ecc73c2a242450a9f1cc858499d468f077ea942867e662c247b5e412 nginx:1.19],SizeBytes:133117205,},ContainerImage{Names:[httpd@sha256:addd70e4ee83f3bc9a4c1c7c41e37927ba47faf639312fc936df3afad7926f5a httpd:2.4.39-alpine],SizeBytes:126894770,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter-amd64@sha256:b63dc612e3cb73f79d2401a4516f794f9f0a83002600ca72e675e41baecff437 directxman12/k8s-prometheus-adapter-amd64:v0.6.0],SizeBytes:53267842,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:25502c57570a2143842478595be9c2a2a3cba2df60b673aef79d6ca80e3eac06 nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44395488,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:9d07c391aeb1a9d02eb4343c113ed01825227c70c32b3cae861711f90191b0fd quay.io/coreos/kube-rbac-proxy:v0.4.1],SizeBytes:41317870,},ContainerImage{Names:[quay.io/coreos/prometheus-operator@sha256:a54e806fb27d2fb0251da4f3b2a3bb5320759af63a54a755788304775f2384a7 quay.io/coreos/prometheus-operator:v0.40.0],SizeBytes:38238457,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[quay.io/coreos/prometheus-config-reloader@sha256:c679a143b24b7731ad1577a9865aa3805426cbf1b25e30807b951dff68466ffd quay.io/coreos/prometheus-config-reloader:v0.40.0],SizeBytes:10131705,},ContainerImage{Names:[jimmidyson/configmap-reload@sha256:d107c7a235c266273b1c3502a391fec374430e5625539403d0de797fa9c556a2 jimmidyson/configmap-reload:v0.3.0],SizeBytes:9700438,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:36553b10a4947067b9fbb7d532951066293a68eae893beba1d9235f7d11a20ad alpine:3.12],SizeBytes:5581415,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:25:54.091: INFO: &#xA;Logging kubelet events for node node1&#xA;May  4 16:25:54.092: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May  4 16:25:54.107: INFO: pod-adoption started at 2021-05-04 16:21:10 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container pod-adoption ready: false, restart count 0&#xA;May  4 16:25:54.107: INFO: collectd-4755t started at 2021-05-04 15:01:51 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: fail-once-local-ltx4r started at 2021-05-04 16:18:38 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container c ready: false, restart count 0&#xA;May  4 16:25:54.107: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-hvrmt started at 2021-05-04 14:52:50 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: liveness-http started at 2021-05-04 15:33:56 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container liveness-http ready: false, restart count 19&#xA;May  4 16:25:54.107: INFO: server-envvars-e2e8d4b8-6525-4f40-9a98-8cccf5c227b4 started at 2021-05-04 16:10:40 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container srv ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: nginx-proxy-node1 started at 2021-05-04 14:51:11 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May  4 16:25:54.107: INFO: kube-proxy-t2mbn started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May  4 16:25:54.107: INFO: cmk-init-discover-node1-m8vvw started at 2021-05-04 14:54:32 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container init ready: false, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container install ready: false, restart count 0&#xA;May  4 16:25:54.107: INFO: busybox-readonly-false-12e191e5-ff02-4f5c-8645-7b0e6b4f0c01 started at 2021-05-04 16:21:18 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container busybox-readonly-false-12e191e5-ff02-4f5c-8645-7b0e6b4f0c01 ready: false, restart count 0&#xA;May  4 16:25:54.107: INFO: kube-multus-ds-amd64-pkmbz started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:25:54.107: INFO: prometheus-k8s-0 started at 2021-05-04 14:56:12 +0000 UTC (0+5 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May  4 16:25:54.107: INFO: &#x9;Container prometheus-config-reloader ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container rules-configmap-reloader ready: true, restart count 0&#xA;May  4 16:25:54.107: INFO: pod-sharedvolume-86936850-dee1-46bf-8b03-52287eae813c started at 2021-05-04 16:22:19 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.107: INFO: &#x9;Container busybox-main-container ready: false, restart count 0&#xA;May  4 16:25:54.107: INFO: &#x9;Container busybox-sub-container ready: false, restart count 0&#xA;May  4 16:25:54.108: INFO: kubernetes-metrics-scraper-678c97765c-6qwqq started at 2021-05-04 14:46:10 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.108: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May  4 16:25:54.108: INFO: cmk-slg76 started at 2021-05-04 14:55:14 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.108: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May  4 16:25:54.108: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May  4 16:25:54.108: INFO: prometheus-operator-5bb8cb9d8f-rrrhf started at 2021-05-04 14:56:03 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.108: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:54.108: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;May  4 16:25:54.108: INFO: node-exporter-k8qd9 started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.108: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:54.108: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:25:54.108: INFO: test-webserver-9b461c4f-7d52-4db1-9027-4951689fb2b4 started at 2021-05-04 16:22:21 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.108: INFO: &#x9;Container test-webserver ready: true, restart count 0&#xA;May  4 16:25:54.108: INFO: kube-flannel-d6pbl started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:25:54.108: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May  4 16:25:54.108: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May  4 16:25:54.108: INFO: node-feature-discovery-worker-wfgl5 started at 2021-05-04 14:51:40 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.108: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;W0504 16:25:54.121108      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:25:54.169: INFO: &#xA;Latency metrics for node node1&#xA;May  4 16:25:54.169: INFO: &#xA;Logging node info for node node2&#xA;May  4 16:25:54.172: INFO: Node Info: &amp;Node{ObjectMeta:{node2   /api/v1/nodes/node2 41567fa7-bb24-4381-9387-e4115195037d 42269 0 2021-05-04 14:44:58 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;32:98:f5:3b:98:5c&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-04 14:44:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:44:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-04 14:52:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-04 14:54:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-04 14:55:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269628928 0} {&lt;nil&gt;} 196552372Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884628480 0} {&lt;nil&gt;} 174692020Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:29 +0000 UTC,LastTransitionTime:2021-05-04 14:47:29 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:53 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:53 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:25:53 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:25:53 +0000 UTC,LastTransitionTime:2021-05-04 14:45:42 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:d47a1c7ab17f44f2ae7ff788700a8d74,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:df3523a1-a74f-4f8b-beb1-29f5ed8699f3,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:511d36b62d758304a5adb26b5996ed37211ab226beb7de4e67994cbecb0279a7 localhost:30500/barometer-collectd:stable],SizeBytes:1464048999,},ContainerImage{Names:[localhost:30500/cmk@sha256:f417461c5e0283b5f2ba8e34dc073a15fe1f9ff6b542330c536c86aa72f7141f localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[nginx@sha256:75a55d33ecc73c2a242450a9f1cc858499d468f077ea942867e662c247b5e412 nginx:1.19],SizeBytes:133117205,},ContainerImage{Names:[httpd@sha256:addd70e4ee83f3bc9a4c1c7c41e37927ba47faf639312fc936df3afad7926f5a httpd:2.4.39-alpine],SizeBytes:126894770,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/sample-apiserver@sha256:ff02aacd9766d597883fabafc7ad604c719a57611db1bcc1564c69a45b000a55 gcr.io/kubernetes-e2e-test-images/sample-apiserver:1.17],SizeBytes:60684726,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:25502c57570a2143842478595be9c2a2a3cba2df60b673aef79d6ca80e3eac06 localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44395488,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:c2fabacbc4e42f3db70f9508e00158b1dce4cf96d91cabaa2eca24e5a0900b66 localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:e074a505e2d62b5119460ab724b2e1df10c8419ef2457f9ce9f3a0f75be3e959 localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[gcr.io/authenticated-image-pulling/alpine@sha256:7ff177862cb50c602bfe81f805969412e619c054a2bbead977d0c276988aa4a0 gcr.io/authenticated-image-pulling/alpine:3.7],SizeBytes:4206620,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:25:54.172: INFO: &#xA;Logging kubelet events for node node2&#xA;May  4 16:25:54.174: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May  4 16:25:54.187: INFO: kubernetes-dashboard-86c6f9df5b-hwbpb started at 2021-05-04 14:46:10 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 1&#xA;May  4 16:25:54.187: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-wt4b2 started at 2021-05-04 14:52:50 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: cmk-2fmbx started at 2021-05-04 14:55:14 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: pod-exec-websocket-2863438c-c2df-4c3c-9cd1-2b53e8002946 started at 2021-05-04 16:22:49 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container main ready: false, restart count 0&#xA;May  4 16:25:54.187: INFO: kube-flannel-lnwkk started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May  4 16:25:54.187: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May  4 16:25:54.187: INFO: cmk-webhook-6c9d5f8578-fr595 started at 2021-05-04 14:55:15 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: ss-0 started at 2021-05-04 16:17:34 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container webserver ready: false, restart count 0&#xA;May  4 16:25:54.187: INFO: liveness-exec started at 2021-05-04 15:33:56 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container liveness-exec ready: false, restart count 6&#xA;May  4 16:25:54.187: INFO: kube-proxy-rfjjf started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May  4 16:25:54.187: INFO: node-feature-discovery-worker-jzjqs started at 2021-05-04 14:51:40 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: node-exporter-5lghf started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: tas-telemetry-aware-scheduling-575ccbc9d4-4nd7x started at 2021-05-04 14:59:02 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:25:54.187: INFO: &#x9;Container tas-controller ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May  4 16:25:54.187: INFO: foo-9dkvq started at 2021-05-04 16:25:24 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container c ready: false, restart count 0&#xA;May  4 16:25:54.188: INFO: kube-multus-ds-amd64-7r2s4 started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:25:54.188: INFO: cmk-init-discover-node2-zlxzj started at 2021-05-04 14:54:52 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May  4 16:25:54.188: INFO: &#x9;Container init ready: false, restart count 0&#xA;May  4 16:25:54.188: INFO: &#x9;Container install ready: false, restart count 0&#xA;May  4 16:25:54.188: INFO: collectd-dhwfp started at 2021-05-04 15:01:51 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May  4 16:25:54.188: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May  4 16:25:54.188: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May  4 16:25:54.188: INFO: fail-once-local-bkr6m started at 2021-05-04 16:18:38 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container c ready: false, restart count 0&#xA;May  4 16:25:54.188: INFO: termination-message-containere34b0019-bcb9-4fa0-9c1b-5eb0017d80c4 started at 2021-05-04 16:22:03 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container termination-message-container ready: false, restart count 0&#xA;May  4 16:25:54.188: INFO: foo-sxtvr started at 2021-05-04 16:25:24 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container c ready: false, restart count 0&#xA;May  4 16:25:54.188: INFO: nginx-proxy-node2 started at 2021-05-04 14:51:11 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:25:54.188: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;W0504 16:25:54.199290      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:25:54.247: INFO: &#xA;Latency metrics for node node2&#xA;May  4 16:25:54.247: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;pods-2302&#34; for this suite.&#xA;</system-out>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume storage capacity unlimited" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Variable Expansion should succeed in writing subpaths in container [sig-storage][Slow] [Conformance]" classname="Kubernetes e2e suite" time="121.481745867">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597&#xA;May  4 16:27:54.342: while waiting for pod to be running&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0002801f0&gt;: {&#xA;        s: &#34;timed out waiting for the condition&#34;,&#xA;    }&#xA;    timed out waiting for the condition&#xA;occurred&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/expansion.go:331</failure>
          <system-out>[BeforeEach] [k8s.io] Variable Expansion&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174&#xA;STEP: Creating a kubernetes client&#xA;May  4 16:25:54.293: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename var-expansion&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[It] should succeed in writing subpaths in container [sig-storage][Slow] [Conformance]&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597&#xA;STEP: creating the pod&#xA;STEP: waiting for pod running&#xA;May  4 16:27:54.342: FAIL: while waiting for pod to be running&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0002801f0&gt;: {&#xA;        s: &#34;timed out waiting for the condition&#34;,&#xA;    }&#xA;    timed out waiting for the condition&#xA;occurred&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/common.glob..func9.8()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/expansion.go:331 +0x4ad&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc001827080)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc001827080)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc001827080, 0x4de37a0)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;[AfterEach] [k8s.io] Variable Expansion&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175&#xA;STEP: Collecting events from namespace &#34;var-expansion-9763&#34;.&#xA;STEP: Found 9 events.&#xA;May  4 16:27:54.347: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: { } Scheduled: Successfully assigned var-expansion-9763/var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720 to node2&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:55 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {multus } AddedInterface: Add eth0 [10.244.3.253/24]&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:55 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:56 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:56 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {kubelet node2} Failed: Error: ErrImagePull&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:57 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {kubelet node2} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:59 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {multus } AddedInterface: Add eth0 [10.244.3.254/24]&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:59 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May  4 16:27:54.347: INFO: At 2021-05-04 16:25:59 +0000 UTC - event for var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May  4 16:27:54.349: INFO: POD                                                 NODE   PHASE    GRACE  CONDITIONS&#xA;May  4 16:27:54.349: INFO: var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:25:54 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:25:54 +0000 UTC ContainersNotReady containers with unready status: [dapi-container]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:25:54 +0000 UTC ContainersNotReady containers with unready status: [dapi-container]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-04 16:25:54 +0000 UTC  }]&#xA;May  4 16:27:54.349: INFO: &#xA;May  4 16:27:54.353: INFO: &#xA;Logging node info for node master1&#xA;May  4 16:27:54.355: INFO: Node Info: &amp;Node{ObjectMeta:{master1   /api/v1/nodes/master1 db982204-549e-4532-90a7-a4410878cfc9 43794 0 2021-05-04 14:43:01 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;3e:f0:43:cb:66:52&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-04 14:43:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:43:02 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-04 14:45:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {nfd-master Update v1 2021-05-04 14:51:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:46 +0000 UTC,LastTransitionTime:2021-05-04 14:47:46 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:01 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:01 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:01 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:47:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:88a0771919594d4187f6704fc7592bf8,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:8e0a253b-2aa4-4467-879e-567e7ba1ffa4,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:c2fabacbc4e42f3db70f9508e00158b1dce4cf96d91cabaa2eca24e5a0900b66 tas-controller:latest localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[nginx@sha256:a97eb9ecc708c8aa715ccfb5e9338f5456e4b65575daf304f108301f3b497314 nginx:1.19.2-alpine],SizeBytes:22052669,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:e074a505e2d62b5119460ab724b2e1df10c8419ef2457f9ce9f3a0f75be3e959 tas-extender:latest localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:27:54.356: INFO: &#xA;Logging kubelet events for node master1&#xA;May  4 16:27:54.358: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May  4 16:27:54.367: INFO: node-feature-discovery-controller-5bf5c49849-72rn6 started at 2021-05-04 14:51:52 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:54.367: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;May  4 16:27:54.367: INFO: kube-flannel-qspzk started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:27:54.367: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May  4 16:27:54.367: INFO: &#x9;Container kube-flannel ready: true, restart count 3&#xA;May  4 16:27:54.367: INFO: kube-multus-ds-amd64-jflvf started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:54.367: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:27:54.367: INFO: coredns-7677f9bb54-qvcd2 started at 2021-05-04 14:46:11 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:54.367: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May  4 16:27:54.367: INFO: node-exporter-jckjs started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:54.367: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:54.367: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:27:54.368: INFO: kube-scheduler-master1 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:54.368: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;May  4 16:27:54.368: INFO: kube-apiserver-master1 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:54.368: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May  4 16:27:54.368: INFO: kube-controller-manager-master1 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:54.368: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May  4 16:27:54.368: INFO: kube-proxy-8j6ch started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:54.368: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May  4 16:27:54.368: INFO: docker-registry-docker-registry-56cbc7bc58-zhf8t started at 2021-05-04 14:48:42 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:54.368: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May  4 16:27:54.368: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;W0504 16:27:54.379807      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:27:55.072: INFO: &#xA;Latency metrics for node master1&#xA;May  4 16:27:55.072: INFO: &#xA;Logging node info for node master2&#xA;May  4 16:27:55.074: INFO: Node Info: &amp;Node{ObjectMeta:{master2   /api/v1/nodes/master2 e2c15170-247b-4e7b-b818-abc807948bf8 43769 0 2021-05-04 14:43:40 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;16:e0:10:a0:e0:62&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-04 14:43:40 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:43:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-04 14:45:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:32 +0000 UTC,LastTransitionTime:2021-05-04 14:47:32 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:40 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:40 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:40 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:47:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:6af568f56589422a9bd68e0270ce0f8c,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:bf27bb77-fad2-4b52-85c3-acb5113fc512,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:27:55.074: INFO: &#xA;Logging kubelet events for node master2&#xA;May  4 16:27:55.076: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May  4 16:27:55.084: INFO: kube-flannel-cxdfr started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May  4 16:27:55.084: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May  4 16:27:55.084: INFO: kube-multus-ds-amd64-dw8tg started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:27:55.084: INFO: dns-autoscaler-5b7b5c9b6f-zbrsq started at 2021-05-04 14:46:08 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;May  4 16:27:55.084: INFO: node-exporter-9c6qf started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:55.084: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:27:55.084: INFO: kube-apiserver-master2 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May  4 16:27:55.084: INFO: kube-controller-manager-master2 started at 2021-05-04 14:47:26 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May  4 16:27:55.084: INFO: kube-scheduler-master2 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;May  4 16:27:55.084: INFO: kube-proxy-6b5t8 started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.084: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;W0504 16:27:55.096052      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:27:55.125: INFO: &#xA;Latency metrics for node master2&#xA;May  4 16:27:55.125: INFO: &#xA;Logging node info for node master3&#xA;May  4 16:27:55.128: INFO: Node Info: &amp;Node{ObjectMeta:{master3   /api/v1/nodes/master3 b533a646-667e-403c-944b-71dec9cc4851 43768 0 2021-05-04 14:43:51 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;16:b0:53:14:f6:c9&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-04 14:43:51 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:43:52 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-04 14:45:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:48:50 +0000 UTC,LastTransitionTime:2021-05-04 14:48:50 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:51 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:51 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:43:51 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:27:50 +0000 UTC,LastTransitionTime:2021-05-04 14:45:46 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:959373dcda56494486f0c2bb0bb496cc,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:0714ca81-c21e-40d6-a288-48d597238e54,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:27:55.128: INFO: &#xA;Logging kubelet events for node master3&#xA;May  4 16:27:55.131: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May  4 16:27:55.139: INFO: kube-apiserver-master3 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May  4 16:27:55.139: INFO: kube-controller-manager-master3 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May  4 16:27:55.139: INFO: kube-scheduler-master3 started at 2021-05-04 14:44:16 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;May  4 16:27:55.139: INFO: kube-proxy-2p5b6 started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May  4 16:27:55.139: INFO: kube-flannel-wznt8 started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May  4 16:27:55.139: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May  4 16:27:55.139: INFO: kube-multus-ds-amd64-cgwz2 started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:27:55.139: INFO: coredns-7677f9bb54-pshfb started at 2021-05-04 14:46:06 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May  4 16:27:55.139: INFO: node-exporter-wvppn started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.139: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:55.139: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;W0504 16:27:55.153083      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:27:55.183: INFO: &#xA;Latency metrics for node master3&#xA;May  4 16:27:55.183: INFO: &#xA;Logging node info for node node1&#xA;May  4 16:27:55.186: INFO: Node Info: &amp;Node{ObjectMeta:{node1   /api/v1/nodes/node1 b8cf7e16-d5c7-4e2c-996a-93d93bd4fa1c 43620 0 2021-05-04 14:44:58 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;e2:50:df:03:d2:13&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-04 14:44:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:44:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-04 14:52:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-04 14:54:35 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-04 14:54:41 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:29 +0000 UTC,LastTransitionTime:2021-05-04 14:47:29 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:46 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:46 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:46 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:27:46 +0000 UTC,LastTransitionTime:2021-05-04 14:47:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:bffc023a4ab84df0b0181bc7b8f509e2,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:dc08af72-abca-4f1d-bd0f-0e8d8eb97de5,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:511d36b62d758304a5adb26b5996ed37211ab226beb7de4e67994cbecb0279a7 localhost:30500/barometer-collectd:stable],SizeBytes:1464048999,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1002569035,},ContainerImage{Names:[opnfv/barometer-collectd@sha256:ed5c574f653e2a39e784ff322033a2319aafde7366c803a88f20f7a2a8bc1efb opnfv/barometer-collectd:stable],SizeBytes:825413035,},ContainerImage{Names:[localhost:30500/cmk@sha256:f417461c5e0283b5f2ba8e34dc073a15fe1f9ff6b542330c536c86aa72f7141f cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:371dc6bf7e0c7ce112a29341b000c40d840aef1dbb4fdcb3ae5c0597e28f3061 golang:alpine3.12],SizeBytes:301097267,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[grafana/grafana@sha256:89304bc2335f4976618548d7b93d165ed67369d3a051d2f627fc4e0aa3d0aff1 grafana/grafana:7.1.0],SizeBytes:179601493,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:d4ba4dd1a9ebb90916d0bfed3c204adcb118ed24546bf8dd2e6b30fc0fd2009e quay.io/prometheus/prometheus:v2.20.0],SizeBytes:144886595,},ContainerImage{Names:[nginx@sha256:75a55d33ecc73c2a242450a9f1cc858499d468f077ea942867e662c247b5e412 nginx:1.19],SizeBytes:133117205,},ContainerImage{Names:[httpd@sha256:addd70e4ee83f3bc9a4c1c7c41e37927ba47faf639312fc936df3afad7926f5a httpd:2.4.39-alpine],SizeBytes:126894770,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter-amd64@sha256:b63dc612e3cb73f79d2401a4516f794f9f0a83002600ca72e675e41baecff437 directxman12/k8s-prometheus-adapter-amd64:v0.6.0],SizeBytes:53267842,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:25502c57570a2143842478595be9c2a2a3cba2df60b673aef79d6ca80e3eac06 nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44395488,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:9d07c391aeb1a9d02eb4343c113ed01825227c70c32b3cae861711f90191b0fd quay.io/coreos/kube-rbac-proxy:v0.4.1],SizeBytes:41317870,},ContainerImage{Names:[quay.io/coreos/prometheus-operator@sha256:a54e806fb27d2fb0251da4f3b2a3bb5320759af63a54a755788304775f2384a7 quay.io/coreos/prometheus-operator:v0.40.0],SizeBytes:38238457,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[quay.io/coreos/prometheus-config-reloader@sha256:c679a143b24b7731ad1577a9865aa3805426cbf1b25e30807b951dff68466ffd quay.io/coreos/prometheus-config-reloader:v0.40.0],SizeBytes:10131705,},ContainerImage{Names:[jimmidyson/configmap-reload@sha256:d107c7a235c266273b1c3502a391fec374430e5625539403d0de797fa9c556a2 jimmidyson/configmap-reload:v0.3.0],SizeBytes:9700438,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:36553b10a4947067b9fbb7d532951066293a68eae893beba1d9235f7d11a20ad alpine:3.12],SizeBytes:5581415,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:27:55.186: INFO: &#xA;Logging kubelet events for node node1&#xA;May  4 16:27:55.189: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May  4 16:27:55.206: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-hvrmt started at 2021-05-04 14:52:50 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: busybox-user-65534-68ff21a5-e8a6-436a-b96d-8961d5f7a8a0 started at 2021-05-04 16:26:47 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container busybox-user-65534-68ff21a5-e8a6-436a-b96d-8961d5f7a8a0 ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: liveness-http started at 2021-05-04 15:33:56 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container liveness-http ready: true, restart count 21&#xA;May  4 16:27:55.207: INFO: dns-test-06270070-c356-4d7e-830e-8ea00f5fb735 started at 2021-05-04 16:27:49 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container jessie-querier ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container querier ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container webserver ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: nginx-proxy-node1 started at 2021-05-04 14:51:11 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May  4 16:27:55.207: INFO: kube-proxy-t2mbn started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May  4 16:27:55.207: INFO: pod-projected-secrets-b8a5b5e3-4702-4eb0-b5de-e380759f490c started at 2021-05-04 16:27:51 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container projected-secret-volume-test ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: cmk-init-discover-node1-m8vvw started at 2021-05-04 14:54:32 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container init ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container install ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: var-expansion-98b73d79-7107-4138-a06e-af820041f2eb started at 2021-05-04 16:27:03 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container dapi-container ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: kube-multus-ds-amd64-pkmbz started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:27:55.207: INFO: simpletest-rc-to-be-deleted-ktdvd started at 2021-05-04 16:27:48 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container nginx ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: prometheus-k8s-0 started at 2021-05-04 14:56:12 +0000 UTC (0+5 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May  4 16:27:55.207: INFO: &#x9;Container prometheus-config-reloader ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container rules-configmap-reloader ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: simpletest-rc-to-be-deleted-fqp76 started at 2021-05-04 16:27:48 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container nginx ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: kubernetes-metrics-scraper-678c97765c-6qwqq started at 2021-05-04 14:46:10 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May  4 16:27:55.207: INFO: cmk-slg76 started at 2021-05-04 14:55:14 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: prometheus-operator-5bb8cb9d8f-rrrhf started at 2021-05-04 14:56:03 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: node-exporter-k8qd9 started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: test-host-network-pod started at 2021-05-04 16:27:39 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container busybox-1 ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container busybox-2 ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: kube-flannel-d6pbl started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May  4 16:27:55.207: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May  4 16:27:55.207: INFO: node-feature-discovery-worker-wfgl5 started at 2021-05-04 14:51:40 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: simpletest-rc-to-be-deleted-bv44b started at 2021-05-04 16:27:48 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container nginx ready: false, restart count 0&#xA;May  4 16:27:55.207: INFO: collectd-4755t started at 2021-05-04 15:01:51 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:55.207: INFO: fail-once-local-ltx4r started at 2021-05-04 16:18:38 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.207: INFO: &#x9;Container c ready: false, restart count 0&#xA;W0504 16:27:55.220912      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:27:55.634: INFO: &#xA;Latency metrics for node node1&#xA;May  4 16:27:55.634: INFO: &#xA;Logging node info for node node2&#xA;May  4 16:27:55.636: INFO: Node Info: &amp;Node{ObjectMeta:{node2   /api/v1/nodes/node2 41567fa7-bb24-4381-9387-e4115195037d 43883 0 2021-05-04 14:44:58 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;32:98:f5:3b:98:5c&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-04 14:44:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-04 14:44:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-04 14:45:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-04 14:52:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-04 14:54:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-04 14:55:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269628928 0} {&lt;nil&gt;} 196552372Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884628480 0} {&lt;nil&gt;} 174692020Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-04 14:47:29 +0000 UTC,LastTransitionTime:2021-05-04 14:47:29 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:53 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:53 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-04 16:27:53 +0000 UTC,LastTransitionTime:2021-05-04 14:44:58 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-04 16:27:53 +0000 UTC,LastTransitionTime:2021-05-04 14:45:42 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:d47a1c7ab17f44f2ae7ff788700a8d74,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:df3523a1-a74f-4f8b-beb1-29f5ed8699f3,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:511d36b62d758304a5adb26b5996ed37211ab226beb7de4e67994cbecb0279a7 localhost:30500/barometer-collectd:stable],SizeBytes:1464048999,},ContainerImage{Names:[localhost:30500/cmk@sha256:f417461c5e0283b5f2ba8e34dc073a15fe1f9ff6b542330c536c86aa72f7141f localhost:30500/cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:726615179,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[nginx@sha256:75a55d33ecc73c2a242450a9f1cc858499d468f077ea942867e662c247b5e412 nginx:1.19],SizeBytes:133117205,},ContainerImage{Names:[httpd@sha256:addd70e4ee83f3bc9a4c1c7c41e37927ba47faf639312fc936df3afad7926f5a httpd:2.4.39-alpine],SizeBytes:126894770,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/sample-apiserver@sha256:ff02aacd9766d597883fabafc7ad604c719a57611db1bcc1564c69a45b000a55 gcr.io/kubernetes-e2e-test-images/sample-apiserver:1.17],SizeBytes:60684726,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:25502c57570a2143842478595be9c2a2a3cba2df60b673aef79d6ca80e3eac06 localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44395488,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:c2fabacbc4e42f3db70f9508e00158b1dce4cf96d91cabaa2eca24e5a0900b66 localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:e074a505e2d62b5119460ab724b2e1df10c8419ef2457f9ce9f3a0f75be3e959 localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[gcr.io/authenticated-image-pulling/alpine@sha256:7ff177862cb50c602bfe81f805969412e619c054a2bbead977d0c276988aa4a0 gcr.io/authenticated-image-pulling/alpine:3.7],SizeBytes:4206620,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May  4 16:27:55.637: INFO: &#xA;Logging kubelet events for node node2&#xA;May  4 16:27:55.639: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May  4 16:27:55.654: INFO: kube-multus-ds-amd64-7r2s4 started at 2021-05-04 14:45:46 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May  4 16:27:55.654: INFO: cmk-init-discover-node2-zlxzj started at 2021-05-04 14:54:52 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container init ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container install ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: collectd-dhwfp started at 2021-05-04 15:01:51 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: fail-once-local-bkr6m started at 2021-05-04 16:18:38 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container c ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: foo-sxtvr started at 2021-05-04 16:25:24 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container c ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: test-pod started at 2021-05-04 16:27:33 +0000 UTC (0+3 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container busybox-1 ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container busybox-2 ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container busybox-3 ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: downward-api-6bd37b02-f45d-43d9-bbeb-bb1f5f8a4813 started at 2021-05-04 16:27:45 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container dapi-container ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: simpletest-rc-to-be-deleted-l674c started at 2021-05-04 16:27:48 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container nginx ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: simpletest-rc-to-be-deleted-bxkqd started at 2021-05-04 16:27:48 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container nginx ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: nginx-proxy-node2 started at 2021-05-04 14:51:11 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May  4 16:27:55.654: INFO: kubernetes-dashboard-86c6f9df5b-hwbpb started at 2021-05-04 14:46:10 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 1&#xA;May  4 16:27:55.654: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-wt4b2 started at 2021-05-04 14:52:50 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: cmk-2fmbx started at 2021-05-04 14:55:14 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: pod-exec-websocket-2863438c-c2df-4c3c-9cd1-2b53e8002946 started at 2021-05-04 16:22:49 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container main ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: pod-service-account-55278b67-f302-4d69-b992-0113a6bbdd84 started at 2021-05-04 16:26:40 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container test ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: kube-flannel-lnwkk started at 2021-05-04 14:45:37 +0000 UTC (1+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May  4 16:27:55.654: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May  4 16:27:55.654: INFO: cmk-webhook-6c9d5f8578-fr595 started at 2021-05-04 14:55:15 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: var-expansion-792ff743-a8c5-4f3a-94b8-4968bd4cf720 started at 2021-05-04 16:25:54 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container dapi-container ready: false, restart count 0&#xA;May  4 16:27:55.654: INFO: liveness-exec started at 2021-05-04 15:33:56 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container liveness-exec ready: false, restart count 6&#xA;May  4 16:27:55.654: INFO: kube-proxy-rfjjf started at 2021-05-04 14:45:01 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May  4 16:27:55.654: INFO: node-feature-discovery-worker-jzjqs started at 2021-05-04 14:51:40 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: node-exporter-5lghf started at 2021-05-04 14:56:10 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: tas-telemetry-aware-scheduling-575ccbc9d4-4nd7x started at 2021-05-04 14:59:02 +0000 UTC (0+2 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container tas-controller ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May  4 16:27:55.654: INFO: foo-9dkvq started at 2021-05-04 16:25:24 +0000 UTC (0+1 container statuses recorded)&#xA;May  4 16:27:55.654: INFO: &#x9;Container c ready: false, restart count 0&#xA;W0504 16:27:55.666186      38 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May  4 16:27:55.765: INFO: &#xA;Latency metrics for node node2&#xA;May  4 16:27:55.765: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;var-expansion-9763&#34; for this suite.&#xA;</system-out>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Character Device [Slow] Should fail on mounting non-existent character device &#39;does-not-exist-char-dev&#39; when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
  </testsuite>