<?xml version="1.0" encoding="UTF-8"?>
  <testsuite name="Kubernetes e2e suite" tests="4" failures="1" errors="0" time="689.606">
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] NodeLease when the NodeLease feature is enabled the kubelet should create and update a lease in the kube-node-lease namespace" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Variable Expansion should allow substituting values in a volume subpath [sig-storage] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Cluster level logging implemented by Stackdriver should ingest logs [Feature:StackdriverLogging]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume FStype [Feature:vsphere] verify fstype - ext3 formatted volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] Multi-AZ Cluster Volumes [sig-storage] should schedule pods in the same zones as statically provisioned PVs" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Directory [Slow] Should fail on mounting non-existent directory &#39;does-not-exist-dir&#39; when HostPathType is HostPathDirectory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV, based on a VSAN capability, datastore and compatible zone specified in storage class" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should allow ingress access from updated namespace [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] NodeProblemDetector [DisabledForLargeClusters] should run without error" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should scale up when non expendable pod is created [Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] RuntimeClass should run a Pod requesting a RuntimeClass with scheduling [NodeFeature:RuntimeHandler] [Disruptive] " classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: gce-localssd-scsi-fs] [Serial] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pod requesting EmptyDir volume is pending [Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cluster-lifecycle] [Feature:BootstrapTokens] should resign the bootstrap tokens when the clusterInfo ConfigMap updated [Serial][Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Directory [Slow] Should be able to mount directory &#39;adir&#39; successfully when HostPathType is HostPathUnset" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap Should fail non-optional pod creation due to the key in the configMap object does not exist [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  Pods sharing a single local PV [Serial] all pods should be running" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Disk Size [Feature:vsphere] verify dynamically provisioned pv has size rounded up correctly" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should allow egress access to server in CIDR block [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] Should scale up GPU pool from 0 [GpuType:] [Feature:ClusterSizeAutoscalingGpu]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking should function for pod-pod: sctp [Feature:SCTPConnectivity]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl client-side validation should create/apply a CR with unknown fields for CRD with no validation schema" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : projected" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DisruptionController evictions: enough pods, replicaSet, percentage =&gt; should allow an eviction" classname="Kubernetes e2e suite" time="26.067412338"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Services should update nodePort: udp [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] One pod requesting one prebound PVC should be able to mount volume and read from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: block] One pod requesting one prebound PVC should be able to mount volume and write from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] RuntimeClass should reject a Pod requesting a RuntimeClass with an unconfigured handler" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Docker Containers should be able to override the image&#39;s default command and arguments [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl get componentstatuses should get componentstatuses" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV, based on the allowed zones, datastore and storage policy specified in storage class" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-bindmounted] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should allow privilege escalation when not explicitly set and uid != 0 [LinuxOnly] [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a persistent volume claim with a storage class. [sig-storage]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should not allow access by TCP when a policy specifies only SCTP [Feature:NetworkPolicy] [Feature:SCTP]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device &#39;ablkdev&#39; when HostPathType is HostPathDirectory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] ReplicationController should serve a basic image on each replica with a public image  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Pod Disks schedule pods each with a PD, delete pod and verify detach [Slow] for read-only PD with pod delete grace period of &#34;default (30s)&#34;" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType File [Slow] Should fail on mounting file &#39;afile&#39; when HostPathType is HostPathBlockDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS configMap nameserver Change stubDomain should be able to change stubDomain configuration [Slow][Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI online volume expansion should expand volume without restarting pod if attach=on, nodeExpansion=on" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV with storage policy specified in storage class in waitForFirstConsumer binding mode" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStackAlphaFeature] [LinuxOnly] should create service with ipv6 cluster ip [Feature:IPv6DualStackAlphaFeature:Phase2]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] CronJob should replace jobs when ReplaceConcurrent" classname="Kubernetes e2e suite" time="122.054713973"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] SSH should SSH to all nodes and run commands" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Regional PD RegionalPD should failover to a different zone when all nodes in one zone become unreachable [Slow] [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] CronJob should schedule multiple jobs concurrently" classname="Kubernetes e2e suite" time="120.047090724"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Pod Disks [Serial] attach on previously attached volumes should work" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should have accelerator metrics [Feature:StackdriverAcceleratorMonitoring]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI NodeStage error cases [Slow] should call NodeUnstage after NodeStage ephemeral error" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] Shouldn&#39;t perform scale up operation and should list unhealthy status if most of the cluster is broken[Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Update Demo should scale a replication controller  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStackAlphaFeature] [LinuxOnly] should have ipv4 and ipv6 internal node ip" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Kubelet [Serial] [Slow] [k8s.io] [sig-node] regular resource usage tracking [Feature:RegularResourceUsageTracking] resource tracking for 100 pods per node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory &#39;adir&#39; when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume limits should verify that all nodes have volume limits" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Events should ensure that an event can be fetched, patched, deleted, and listed [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl expose should create services for rc  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] kubelet [k8s.io] [sig-node] Clean up pods on node kubelet should be able to delete 10 pods per node in 1m0s." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] Should scale down GPU pool from 1 [GpuType:] [Feature:ClusterSizeAutoscalingGpu]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device &#39;ablkdev&#39; when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should support a &#39;default-deny-ingress&#39; policy [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] EndpointSlice should have Endpoints and EndpointSlices pointing to API Server" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] Security Context should support seccomp alpha runtime/default annotation [Feature:Seccomp] [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] [HPA] Horizontal pod autoscaling (scale resource: Custom Metrics from Stackdriver) should scale down with Custom Metric of type Pod from Stackdriver with Prometheus [Feature:CustomMetricsAutoscaling]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-ui] Kubernetes Dashboard [Feature:Dashboard] should check that the kubernetes-dashboard instance is alive" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  StatefulSet with pod affinity [Slow] should use volumes on one node when pod has affinity" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DaemonRestart [Disruptive] Kubelet should not restart containers across restart" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-node] PreStop should call prestop when killing a pod  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl label should update the label on a resource  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] [sig-cloud-provider-gcp] Upgrade [Feature:Upgrade] node upgrade should maintain a functioning cluster [Feature:NodeUpgrade]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify a pod is created and attached to a dynamically created PV with storage policy specified in storage class in waitForFirstConsumer binding mode with allowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning [k8s.io] GlusterDynamicProvisioner should create and delete persistent volumes [fast]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) [Feature:GenericEphemeralVolume] (immediate-binding)] ephemeral should create read-only inline ephemeral volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] StatefulSet [k8s.io] Deploy clustered applications [Feature:StatefulSet] [Slow] should creating a working CockroachDB cluster" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] CronJob should not emit unexpected warnings" classname="Kubernetes e2e suite" time="420.373238974">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:174&#xA;May 12 19:48:02.653: Failed to ensure at least on finished job exists in namespace cronjob-6462&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0002ba200&gt;: {&#xA;        s: &#34;timed out waiting for the condition&#34;,&#xA;    }&#xA;    timed out waiting for the condition&#xA;occurred&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:185</failure>
          <system-out>[BeforeEach] [sig-apps] CronJob&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174&#xA;STEP: Creating a kubernetes client&#xA;May 12 19:41:02.612: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename cronjob&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[BeforeEach] [sig-apps] CronJob&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58&#xA;[It] should not emit unexpected warnings&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:174&#xA;STEP: Creating a cronjob&#xA;STEP: Ensuring at least two jobs and at least one finished job exists by listing jobs explicitly&#xA;May 12 19:48:02.653: FAIL: Failed to ensure at least on finished job exists in namespace cronjob-6462&#xA;Unexpected error:&#xA;    &lt;*errors.errorString | 0xc0002ba200&gt;: {&#xA;        s: &#34;timed out waiting for the condition&#34;,&#xA;    }&#xA;    timed out waiting for the condition&#xA;occurred&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/apps.glob..func1.6()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:185 +0x40d&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc000681980)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc000681980)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc000681980, 0x4de37a0)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;[AfterEach] [sig-apps] CronJob&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175&#xA;STEP: Collecting events from namespace &#34;cronjob-6462&#34;.&#xA;STEP: Found 60 events.&#xA;May 12 19:48:02.658: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for concurrent-1620848520-6rh26: { } Scheduled: Successfully assigned cronjob-6462/concurrent-1620848520-6rh26 to node1&#xA;May 12 19:48:02.658: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for concurrent-1620848580-b6r88: { } Scheduled: Successfully assigned cronjob-6462/concurrent-1620848580-b6r88 to node2&#xA;May 12 19:48:02.658: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for concurrent-1620848640-b24dv: { } Scheduled: Successfully assigned cronjob-6462/concurrent-1620848640-b24dv to node2&#xA;May 12 19:48:02.658: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for concurrent-1620848700-56mr7: { } Scheduled: Successfully assigned cronjob-6462/concurrent-1620848700-56mr7 to node2&#xA;May 12 19:48:02.658: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for concurrent-1620848760-d6r4m: { } Scheduled: Successfully assigned cronjob-6462/concurrent-1620848760-d6r4m to node2&#xA;May 12 19:48:02.658: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for concurrent-1620848820-q6p7s: { } Scheduled: Successfully assigned cronjob-6462/concurrent-1620848820-q6p7s to node2&#xA;May 12 19:48:02.658: INFO: At 2021-05-12 19:42:02 +0000 UTC - event for concurrent: {cronjob-controller } SuccessfulCreate: Created job concurrent-1620848520&#xA;May 12 19:48:02.658: INFO: At 2021-05-12 19:42:02 +0000 UTC - event for concurrent-1620848520: {job-controller } SuccessfulCreate: Created pod: concurrent-1620848520-6rh26&#xA;May 12 19:48:02.658: INFO: At 2021-05-12 19:42:05 +0000 UTC - event for concurrent-1620848520-6rh26: {kubelet node1} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:42:05 +0000 UTC - event for concurrent-1620848520-6rh26: {multus } AddedInterface: Add eth0 [10.244.3.17/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:42:06 +0000 UTC - event for concurrent-1620848520-6rh26: {kubelet node1} Failed: Error: ErrImagePull&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:42:06 +0000 UTC - event for concurrent-1620848520-6rh26: {kubelet node1} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:42:07 +0000 UTC - event for concurrent-1620848520-6rh26: {kubelet node1} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:42:09 +0000 UTC - event for concurrent-1620848520-6rh26: {multus } AddedInterface: Add eth0 [10.244.3.18/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:42:09 +0000 UTC - event for concurrent-1620848520-6rh26: {kubelet node1} Failed: Error: ImagePullBackOff&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:42:09 +0000 UTC - event for concurrent-1620848520-6rh26: {kubelet node1} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:02 +0000 UTC - event for concurrent: {cronjob-controller } SuccessfulCreate: Created job concurrent-1620848580&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:02 +0000 UTC - event for concurrent-1620848580: {job-controller } SuccessfulCreate: Created pod: concurrent-1620848580-b6r88&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:03 +0000 UTC - event for concurrent-1620848580-b6r88: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:03 +0000 UTC - event for concurrent-1620848580-b6r88: {multus } AddedInterface: Add eth0 [10.244.4.95/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:04 +0000 UTC - event for concurrent-1620848580-b6r88: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:04 +0000 UTC - event for concurrent-1620848580-b6r88: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:05 +0000 UTC - event for concurrent-1620848580-b6r88: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:43:05 +0000 UTC - event for concurrent-1620848580-b6r88: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:02 +0000 UTC - event for concurrent: {cronjob-controller } SuccessfulCreate: Created job concurrent-1620848640&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:02 +0000 UTC - event for concurrent-1620848640: {job-controller } SuccessfulCreate: Created pod: concurrent-1620848640-b24dv&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:03 +0000 UTC - event for concurrent-1620848640-b24dv: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:03 +0000 UTC - event for concurrent-1620848640-b24dv: {multus } AddedInterface: Add eth0 [10.244.4.96/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:04 +0000 UTC - event for concurrent-1620848640-b24dv: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:04 +0000 UTC - event for concurrent-1620848640-b24dv: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:05 +0000 UTC - event for concurrent-1620848640-b24dv: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:44:05 +0000 UTC - event for concurrent-1620848640-b24dv: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:02 +0000 UTC - event for concurrent: {cronjob-controller } SuccessfulCreate: Created job concurrent-1620848700&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:02 +0000 UTC - event for concurrent-1620848700: {job-controller } SuccessfulCreate: Created pod: concurrent-1620848700-56mr7&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:04 +0000 UTC - event for concurrent-1620848700-56mr7: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:04 +0000 UTC - event for concurrent-1620848700-56mr7: {multus } AddedInterface: Add eth0 [10.244.4.97/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:05 +0000 UTC - event for concurrent-1620848700-56mr7: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:05 +0000 UTC - event for concurrent-1620848700-56mr7: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:05 +0000 UTC - event for concurrent-1620848700-56mr7: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:45:05 +0000 UTC - event for concurrent-1620848700-56mr7: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:02 +0000 UTC - event for concurrent: {cronjob-controller } SuccessfulCreate: Created job concurrent-1620848760&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:02 +0000 UTC - event for concurrent-1620848760: {job-controller } SuccessfulCreate: Created pod: concurrent-1620848760-d6r4m&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:03 +0000 UTC - event for concurrent-1620848760-d6r4m: {multus } AddedInterface: Add eth0 [10.244.4.98/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:04 +0000 UTC - event for concurrent-1620848760-d6r4m: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:05 +0000 UTC - event for concurrent-1620848760-d6r4m: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:05 +0000 UTC - event for concurrent-1620848760-d6r4m: {kubelet node2} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:05 +0000 UTC - event for concurrent-1620848760-d6r4m: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:07 +0000 UTC - event for concurrent-1620848760-d6r4m: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:07 +0000 UTC - event for concurrent-1620848760-d6r4m: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:46:07 +0000 UTC - event for concurrent-1620848760-d6r4m: {multus } AddedInterface: Add eth0 [10.244.4.99/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:02 +0000 UTC - event for concurrent: {cronjob-controller } SuccessfulCreate: Created job concurrent-1620848820&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:02 +0000 UTC - event for concurrent-1620848820: {job-controller } SuccessfulCreate: Created pod: concurrent-1620848820-q6p7s&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:04 +0000 UTC - event for concurrent-1620848820-q6p7s: {kubelet node2} Pulling: Pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:04 +0000 UTC - event for concurrent-1620848820-q6p7s: {multus } AddedInterface: Add eth0 [10.244.4.100/24]&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:05 +0000 UTC - event for concurrent-1620848820-q6p7s: {kubelet node2} Failed: Error: ErrImagePull&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:05 +0000 UTC - event for concurrent-1620848820-q6p7s: {kubelet node2} Failed: Failed to pull image &#34;docker.io/library/busybox:1.29&#34;: rpc error: code = Unknown desc = Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:06 +0000 UTC - event for concurrent-1620848820-q6p7s: {kubelet node2} SandboxChanged: Pod sandbox changed, it will be killed and re-created.&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:07 +0000 UTC - event for concurrent-1620848820-q6p7s: {kubelet node2} BackOff: Back-off pulling image &#34;docker.io/library/busybox:1.29&#34;&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:07 +0000 UTC - event for concurrent-1620848820-q6p7s: {kubelet node2} Failed: Error: ImagePullBackOff&#xA;May 12 19:48:02.659: INFO: At 2021-05-12 19:47:07 +0000 UTC - event for concurrent-1620848820-q6p7s: {multus } AddedInterface: Add eth0 [10.244.4.101/24]&#xA;May 12 19:48:02.662: INFO: POD                          NODE   PHASE    GRACE  CONDITIONS&#xA;May 12 19:48:02.662: INFO: concurrent-1620848520-6rh26  node1  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:02 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:42:02 +0000 UTC  }]&#xA;May 12 19:48:02.662: INFO: concurrent-1620848580-b6r88  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:43:02 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:43:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:43:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:43:02 +0000 UTC  }]&#xA;May 12 19:48:02.662: INFO: concurrent-1620848640-b24dv  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:44:02 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:44:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:44:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:44:02 +0000 UTC  }]&#xA;May 12 19:48:02.662: INFO: concurrent-1620848700-56mr7  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:45:02 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:45:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:45:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:45:02 +0000 UTC  }]&#xA;May 12 19:48:02.662: INFO: concurrent-1620848760-d6r4m  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:46:02 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:46:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:46:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:46:02 +0000 UTC  }]&#xA;May 12 19:48:02.662: INFO: concurrent-1620848820-q6p7s  node2  Pending         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:47:02 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:47:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:47:02 +0000 UTC ContainersNotReady containers with unready status: [c]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-12 19:47:02 +0000 UTC  }]&#xA;May 12 19:48:02.662: INFO: &#xA;May 12 19:48:02.667: INFO: &#xA;Logging node info for node master1&#xA;May 12 19:48:02.670: INFO: Node Info: &amp;Node{ObjectMeta:{master1   /api/v1/nodes/master1 03e64d13-444c-41eb-b6bd-3745f05cd1cd 73451 0 2021-05-12 16:30:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;ce:94:4b:fc:cf:bc&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:30:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:30:45 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {kube-controller-manager Update v1 2021-05-12 16:31:10 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:27 +0000 UTC,LastTransitionTime:2021-05-12 16:35:27 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:59 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:59 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:59 +0000 UTC,LastTransitionTime:2021-05-12 16:30:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:47:59 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:201a06a0695b47fdbdd7df0fdd94f4dc,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:a7dddfd4-eda7-4a08-9750-d4a3c87c0cdd,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:60f4b5001bb5e7280fddf9143d3ed9bcde4e8016eef54522b5aea6bac9d9774b tas-controller:latest localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[nginx@sha256:a97eb9ecc708c8aa715ccfb5e9338f5456e4b65575daf304f108301f3b497314 nginx:1.19.2-alpine],SizeBytes:22052669,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:1935bf73835eb6d4446668a5484bb3724d97d926b26b41c7bf064aa3a5a8bc5f tas-extender:latest localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:48:02.670: INFO: &#xA;Logging kubelet events for node master1&#xA;May 12 19:48:02.673: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May 12 19:48:02.694: INFO: docker-registry-docker-registry-56cbc7bc58-lfjnt started at 2021-05-12 16:36:16 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May 12 19:48:02.694: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;May 12 19:48:02.694: INFO: kube-apiserver-master1 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:48:02.694: INFO: kube-scheduler-master1 started at 2021-05-12 16:32:00 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;May 12 19:48:02.694: INFO: kube-proxy-v5zpq started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:48:02.694: INFO: kube-flannel-rc2gh started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:48:02.694: INFO: &#x9;Container kube-flannel ready: true, restart count 3&#xA;May 12 19:48:02.694: INFO: coredns-7677f9bb54-4lkxs started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container coredns ready: true, restart count 2&#xA;May 12 19:48:02.694: INFO: kube-controller-manager-master1 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:48:02.694: INFO: kube-multus-ds-amd64-rcwwm started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:48:02.694: INFO: node-exporter-wd9j2 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.694: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.694: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;W0512 19:48:02.709163      31 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:48:02.740: INFO: &#xA;Latency metrics for node master1&#xA;May 12 19:48:02.740: INFO: &#xA;Logging node info for node master2&#xA;May 12 19:48:02.742: INFO: Node Info: &amp;Node{ObjectMeta:{master2   /api/v1/nodes/master2 f3d8b0a9-b7ca-4942-bfc7-69b182fddff8 73443 0 2021-05-12 16:31:23 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;32:9d:92:dd:24:1d&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:31:23 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:31:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-12 16:33:30 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:36:36 +0000 UTC,LastTransitionTime:2021-05-12 16:36:36 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:57 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:57 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:57 +0000 UTC,LastTransitionTime:2021-05-12 16:31:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:47:57 +0000 UTC,LastTransitionTime:2021-05-12 16:33:30 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:ef3726c355844802be21d744dd831d84,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:6feec048-83de-4ff4-ae36-12985bcda218,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:48:02.742: INFO: &#xA;Logging kubelet events for node master2&#xA;May 12 19:48:02.745: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May 12 19:48:02.751: INFO: node-exporter-dprww started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.751: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.751: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:48:02.751: INFO: kube-apiserver-master2 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.751: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:48:02.751: INFO: kube-controller-manager-master2 started at 2021-05-12 16:38:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.751: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:48:02.751: INFO: kube-scheduler-master2 started at 2021-05-12 16:32:00 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.752: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;May 12 19:48:02.752: INFO: kube-proxy-6ljsd started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.752: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:48:02.752: INFO: kube-flannel-pppch started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:48:02.752: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:48:02.752: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 12 19:48:02.752: INFO: kube-multus-ds-amd64-4swdk started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.752: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:48:02.752: INFO: coredns-7677f9bb54-b628b started at 2021-05-12 16:33:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.752: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;W0512 19:48:02.764109      31 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:48:02.788: INFO: &#xA;Latency metrics for node master2&#xA;May 12 19:48:02.788: INFO: &#xA;Logging node info for node master3&#xA;May 12 19:48:02.790: INFO: Node Info: &amp;Node{ObjectMeta:{master3   /api/v1/nodes/master3 a0124635-fd55-4774-af26-32765b659025 73444 0 2021-05-12 16:31:34 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;8e:91:ce:de:d3:10&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-12 16:31:34 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:31:35 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-12 16:33:30 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {nfd-master Update v1 2021-05-12 16:39:03 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:42 +0000 UTC,LastTransitionTime:2021-05-12 16:35:42 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:58 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:58 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:58 +0000 UTC,LastTransitionTime:2021-05-12 16:31:34 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:47:58 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:8966dda31ac243c3b9142da7eb1a3315,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:29bf94f6-9eaa-4435-8e2a-f17b4721b8df,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:9d07c391aeb1a9d02eb4343c113ed01825227c70c32b3cae861711f90191b0fd quay.io/coreos/kube-rbac-proxy:v0.4.1],SizeBytes:41317870,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[quay.io/coreos/prometheus-operator@sha256:a54e806fb27d2fb0251da4f3b2a3bb5320759af63a54a755788304775f2384a7 quay.io/coreos/prometheus-operator:v0.40.0],SizeBytes:38238457,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:48:02.791: INFO: &#xA;Logging kubelet events for node master3&#xA;May 12 19:48:02.793: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May 12 19:48:02.802: INFO: prometheus-operator-5bb8cb9d8f-sqgl4 started at 2021-05-12 16:42:55 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.802: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.802: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;May 12 19:48:02.802: INFO: kube-apiserver-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.802: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 12 19:48:02.802: INFO: kube-controller-manager-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.802: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 12 19:48:02.803: INFO: kube-multus-ds-amd64-4hlf8 started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.803: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:48:02.803: INFO: dns-autoscaler-5b7b5c9b6f-8fzpk started at 2021-05-12 16:33:51 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.803: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;May 12 19:48:02.803: INFO: node-feature-discovery-controller-5bf5c49849-gdz9z started at 2021-05-12 16:38:58 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.803: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;May 12 19:48:02.803: INFO: node-exporter-jzp84 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.803: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.803: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:48:02.803: INFO: kube-scheduler-master3 started at 2021-05-12 16:35:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.803: INFO: &#x9;Container kube-scheduler ready: true, restart count 3&#xA;May 12 19:48:02.803: INFO: kube-proxy-d57dm started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.803: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 12 19:48:02.803: INFO: kube-flannel-hmbzn started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:48:02.803: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:48:02.803: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;W0512 19:48:02.817747      31 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:48:02.840: INFO: &#xA;Latency metrics for node master3&#xA;May 12 19:48:02.840: INFO: &#xA;Logging node info for node node1&#xA;May 12 19:48:02.845: INFO: Node Info: &amp;Node{ObjectMeta:{node1   /api/v1/nodes/node1 2f7b1644-6865-412d-b423-194903813633 73457 0 2021-05-12 16:32:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;6a:b5:62:81:d9:8e&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:25 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-12 16:39:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-12 16:41:28 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:02:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:36:30 +0000 UTC,LastTransitionTime:2021-05-12 16:36:30 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:48:00 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:48:00 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:48:00 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:48:00 +0000 UTC,LastTransitionTime:2021-05-12 16:33:25 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:d9a9ef88c91340689744923e9951f78a,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:d2f244db-a37b-4dba-a8f9-6ea24c6cdb4e,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:96570ae3225d359c9667e2cbd57987388aa9d3be9d5a198ed9b677c7f9f4e450 localhost:30500/barometer-collectd:stable],SizeBytes:1464260582,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1002488025,},ContainerImage{Names:[opnfv/barometer-collectd@sha256:ed5c574f653e2a39e784ff322033a2319aafde7366c803a88f20f7a2a8bc1efb opnfv/barometer-collectd:stable],SizeBytes:825413035,},ContainerImage{Names:[localhost:30500/cmk@sha256:f178577724ef3b8118ca58277a27d594257de6c55a813b61e421df4008a7f73b cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:1636899c10870ab66c48d960a9df620f4f9e86a0c72fbacf36032d27404e7e6c golang:alpine3.12],SizeBytes:301156062,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[grafana/grafana@sha256:89304bc2335f4976618548d7b93d165ed67369d3a051d2f627fc4e0aa3d0aff1 grafana/grafana:7.1.0],SizeBytes:179601493,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:d4ba4dd1a9ebb90916d0bfed3c204adcb118ed24546bf8dd2e6b30fc0fd2009e quay.io/prometheus/prometheus:v2.20.0],SizeBytes:144886595,},ContainerImage{Names:[nginx@sha256:34ac17eb150f557f6ce94ea34e7c03f3899b717307cad633213a9488a431179f nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter-amd64@sha256:b63dc612e3cb73f79d2401a4516f794f9f0a83002600ca72e675e41baecff437 directxman12/k8s-prometheus-adapter-amd64:v0.6.0],SizeBytes:53267842,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:155361826d160d9f566aede3ea35b34f1e0c6422720285b1c22467c2b21a90aa nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392919,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[quay.io/coreos/prometheus-config-reloader@sha256:c679a143b24b7731ad1577a9865aa3805426cbf1b25e30807b951dff68466ffd quay.io/coreos/prometheus-config-reloader:v0.40.0],SizeBytes:10131705,},ContainerImage{Names:[jimmidyson/configmap-reload@sha256:d107c7a235c266273b1c3502a391fec374430e5625539403d0de797fa9c556a2 jimmidyson/configmap-reload:v0.3.0],SizeBytes:9700438,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:36553b10a4947067b9fbb7d532951066293a68eae893beba1d9235f7d11a20ad alpine:3.12],SizeBytes:5581415,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:48:02.846: INFO: &#xA;Logging kubelet events for node node1&#xA;May 12 19:48:02.848: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May 12 19:48:02.862: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-46jff started at 2021-05-12 16:39:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.862: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 12 19:48:02.862: INFO: cmk-init-discover-node1-2x2zk started at 2021-05-12 16:41:25 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: concurrent-1620848880-xpng5 started at 2021-05-12 19:48:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: fail-once-non-local-2vxhb started at 2021-05-12 19:42:09 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: nginx-proxy-node1 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 12 19:48:02.863: INFO: node-feature-discovery-worker-qtn84 started at 2021-05-12 16:38:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: all-succeed-x6rzf started at 2021-05-12 19:36:55 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: concurrent-1620848520-6rh26 started at 2021-05-12 19:42:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: kubernetes-dashboard-86c6f9df5b-vkvbq started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 1&#xA;May 12 19:48:02.863: INFO: prometheus-k8s-0 started at 2021-05-12 16:43:20 +0000 UTC (0+5 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May 12 19:48:02.863: INFO: &#x9;Container prometheus-config-reloader ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container rules-configmap-reloader ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: collectd-5mpmz started at 2021-05-12 16:49:38 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: kube-proxy-r9vsx started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 12 19:48:02.863: INFO: node-exporter-ddxbd started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: fail-once-non-local-dl6pl started at 2021-05-12 19:42:09 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.863: INFO: kube-flannel-r7w6z started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Init container install-cni ready: true, restart count 1&#xA;May 12 19:48:02.863: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 12 19:48:02.863: INFO: kube-multus-ds-amd64-fhzwc started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:48:02.863: INFO: kubernetes-metrics-scraper-678c97765c-s4sgj started at 2021-05-12 16:33:53 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May 12 19:48:02.863: INFO: cmk-v4qwz started at 2021-05-12 16:42:07 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.863: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 12 19:48:02.863: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;W0512 19:48:02.874605      31 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:48:02.904: INFO: &#xA;Latency metrics for node node1&#xA;May 12 19:48:02.904: INFO: &#xA;Logging node info for node node2&#xA;May 12 19:48:02.907: INFO: Node Info: &amp;Node{ObjectMeta:{node2   /api/v1/nodes/node2 16e4ef1f-0c51-44d5-a257-2eaa348c0d52 73428 0 2021-05-12 16:32:43 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;52:9a:ee:23:4c:ac&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-12 16:32:43 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-12 16:33:25 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-12 16:39:06 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-12 16:41:50 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {e2e.test Update v1 2021-05-12 19:08:41 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:example.com/fakecpu&#34;:{}}}}} {kubelet Update v1 2021-05-12 19:08:51 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:example.com/fakecpu&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-12 16:35:14 +0000 UTC,LastTransitionTime:2021-05-12 16:35:14 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:53 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:53 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-12 19:47:53 +0000 UTC,LastTransitionTime:2021-05-12 16:32:43 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-12 19:47:53 +0000 UTC,LastTransitionTime:2021-05-12 16:35:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:eebaf4858fee4a739009f2f7f2717953,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:15bd8119-98f5-4e13-a2fa-897c2224dd6e,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:96570ae3225d359c9667e2cbd57987388aa9d3be9d5a198ed9b677c7f9f4e450 localhost:30500/barometer-collectd:stable],SizeBytes:1464260582,},ContainerImage{Names:[localhost:30500/cmk@sha256:f178577724ef3b8118ca58277a27d594257de6c55a813b61e421df4008a7f73b localhost:30500/cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:726657349,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[nginx@sha256:34ac17eb150f557f6ce94ea34e7c03f3899b717307cad633213a9488a431179f nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/sample-apiserver@sha256:ff02aacd9766d597883fabafc7ad604c719a57611db1bcc1564c69a45b000a55 gcr.io/kubernetes-e2e-test-images/sample-apiserver:1.17],SizeBytes:60684726,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:155361826d160d9f566aede3ea35b34f1e0c6422720285b1c22467c2b21a90aa localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392919,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:60f4b5001bb5e7280fddf9143d3ed9bcde4e8016eef54522b5aea6bac9d9774b localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:1935bf73835eb6d4446668a5484bb3724d97d926b26b41c7bf064aa3a5a8bc5f localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[gcr.io/authenticated-image-pulling/alpine@sha256:7ff177862cb50c602bfe81f805969412e619c054a2bbead977d0c276988aa4a0 gcr.io/authenticated-image-pulling/alpine:3.7],SizeBytes:4206620,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 12 19:48:02.907: INFO: &#xA;Logging kubelet events for node node2&#xA;May 12 19:48:02.910: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May 12 19:48:02.924: INFO: concurrent-1620848580-b6r88 started at 2021-05-12 19:43:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: concurrent-1620848820-q6p7s started at 2021-05-12 19:47:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: all-succeed-rbrzg started at 2021-05-12 19:36:55 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: kube-proxy-grtqc started at 2021-05-12 16:32:45 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 12 19:48:02.924: INFO: cmk-gbw5d started at 2021-05-12 16:42:08 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: backofflimit-vrd8v started at 2021-05-12 19:36:52 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: collectd-tng6x started at 2021-05-12 16:49:38 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: all-pods-removed-mzshw started at 2021-05-12 19:37:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: all-pods-removed-r4vfz started at 2021-05-12 19:37:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: concurrent-1620848640-b24dv started at 2021-05-12 19:44:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: concurrent-1620848700-56mr7 started at 2021-05-12 19:45:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container c ready: false, restart count 0&#xA;May 12 19:48:02.924: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-j6r4r started at 2021-05-12 16:39:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: node-exporter-h5rv7 started at 2021-05-12 16:43:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: kube-multus-ds-amd64-k28rf started at 2021-05-12 16:33:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 12 19:48:02.924: INFO: node-feature-discovery-worker-zjvzk started at 2021-05-12 16:38:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: tas-telemetry-aware-scheduling-575ccbc9d4-8xc75 started at 2021-05-12 16:46:05 +0000 UTC (0+2 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container tas-controller ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: cmk-webhook-6c9d5f8578-fgcvr started at 2021-05-12 16:42:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May 12 19:48:02.924: INFO: nginx-proxy-node2 started at 2021-05-12 16:38:15 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 12 19:48:02.924: INFO: kube-flannel-rqtcs started at 2021-05-12 16:33:20 +0000 UTC (1+1 container statuses recorded)&#xA;May 12 19:48:02.924: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 12 19:48:02.925: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 12 19:48:02.925: INFO: cmk-init-discover-node2-qrd9v started at 2021-05-12 16:41:44 +0000 UTC (0+3 container statuses recorded)&#xA;May 12 19:48:02.925: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 12 19:48:02.925: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 12 19:48:02.925: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 12 19:48:02.925: INFO: concurrent-1620848760-d6r4m started at 2021-05-12 19:46:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 12 19:48:02.925: INFO: &#x9;Container c ready: false, restart count 0&#xA;W0512 19:48:02.943852      31 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 12 19:48:02.976: INFO: &#xA;Latency metrics for node node2&#xA;May 12 19:48:02.976: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;cronjob-6462&#34; for this suite.&#xA;</system-out>
      </testcase>
  </testsuite>