<?xml version="1.0" encoding="UTF-8"?>
  <testsuite name="Kubernetes e2e suite" tests="1" failures="1" errors="0" time="240.534">
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Sysctls [LinuxOnly] [NodeFeature:Sysctls] should support sysctls [MinimumKubeletVersion:1.21] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with different fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-scheduling] SchedulerPredicates [Serial] validates local ephemeral storage resource limits of pods that are allowed to run [Feature:LocalStorageCapacityIsolation]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Pod Disks detach in a disrupted environment [Slow] [Disruptive] when node&#39;s API object is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaler scalability [Slow] should scale up at all [Feature:ClusterAutoscalerScalability1]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read-only inline ephemeral volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client should deny ingress access to updated pod [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with different fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning DynamicProvisioner Default should be disabled by removing the default annotation [Serial] [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] DNS horizontal autoscaling kube-dns-autoscaler should scale kube-dns pods in both nonfaulty and faulty scenarios" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Network Partition [Disruptive] [Slow] Pods should be evicted from unready Node [Feature:TaintEviction] All pods on the unreachable node should be marked as NotReady upon the node turn NotReady AND all pods should be evicted after eviction timeout passes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup skips ownership changes to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPath should support subPath [NodeConformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client should deny ingress from pods on other namespaces [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce multiple, stacked policies with overlapping podSelectors [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSIStorageCapacity CSIStorageCapacity disabled" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV and a pre-bound PVC: test write access" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI workload information using mock driver contain ephemeral=true when using inline volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] API priority and fairness should ensure that requests can&#39;t be drowned out (priority)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected secret Should fail non-optional pod creation due to secret object does not exist [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] files with FSGroup ownership should support (root,0644,tmpfs)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Stackdriver Metadata Agent [Feature:StackdriverMetadataAgent]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Storage Policy Based Volume Provisioning [Feature:vsphere] verify VSAN storage capability with invalid diskStripes value is not honored for dynamically provisioned pvc using storageclass" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Firewall rule should have correct firewall rules for e2e cluster" classname="Kubernetes e2e suite" time="0.036352392">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] stateful Upgrade [Feature:StatefulUpgrade] stateful upgrade should maintain a functioning cluster" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Deployment should run the lifecycle of a Deployment [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] [HPA] Horizontal pod autoscaling (scale resource: Custom Metrics from Stackdriver) should scale down with External Metric with target average value from Stackdriver [Feature:CustomMetricsAutoscaling]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cloud-provider-gcp] Nodes [Disruptive] Resize [Slow] should be able to delete nodes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] StatefulSet Deploy clustered applications [Feature:StatefulSet] [Slow] should creating a working redis cluster" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] [Feature:Windows] Windows volume mounts  check volume mount permissions container should have readOnly permissions on hostMapPath" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Custom Metrics - Stackdriver Adapter for external metrics [Feature:StackdriverExternalMetrics]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] ESIPP [Slow] should work for type=NodePort" classname="Kubernetes e2e suite" time="0.039781243">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Zone Support [Feature:vsphere] Verify PVC creation with invalid zone specified in storage class fails" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client should deny egress from pods based on PodSelector [Feature:NetworkPolicy] " classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-instrumentation] Stackdriver Monitoring should run Custom Metrics - Stackdriver Adapter for new resource model [Feature:StackdriverCustomMetrics]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Probing container should be restarted with a exec &#34;cat /tmp/health&#34; liveness probe [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] IngressClass API  should support creating IngressClass API operations [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl create quota should reject quota with invalid scopes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota&#39;s priority class scope (quota set to pod count: 1) against 2 pods with same priority class." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] shouldn&#39;t scale down when non expendable pod is running [Feature:ClusterSizeAutoscalingScaleDown]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] [Feature:Windows] SecurityContext should ignore Linux Specific SecurityContext if set" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volume-lifecycle-performance should provision volumes at scale within performance constraints [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI workload information using mock driver should not be passed when CSIDriver does not exist" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] [Feature:Windows] Memory Limits [Serial] [Slow] attempt to deploy past allocatable memory limits should fail deployments of pods once there isn&#39;t enough memory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Update Demo should scale a replication controller  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should provide container&#39;s cpu limit [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] ESIPP [Slow] should work from pods" classname="Kubernetes e2e suite" time="0.031933655">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] [Feature:Example] Liveness liveness pods should be automatically restarted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Servers with support for API chunking should return chunks of results for list calls" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup skips ownership changes to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] EndpointSlice should support creating EndpointSlice API operations [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Services should function for node-Service: http" classname="Kubernetes e2e suite" time="124.629698604">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/networking.go:198&#xA;May 20 23:24:38.909: failed dialing endpoint, failed to find expected endpoints, &#xA;tries 34&#xA;Command curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName&#xA;retrieved map[]&#xA;expected map[netserver-0:{} netserver-1:{}]&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/internal/leafnodes/runner.go:113</failure>
          <system-out>[BeforeEach] [sig-network] Networking&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185&#xA;STEP: Creating a kubernetes client&#xA;May 20 23:22:35.806: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename nettest&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[BeforeEach] [sig-network] Networking&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/networking.go:83&#xA;STEP: Executing a successful http request from the external internet&#xA;[It] should function for node-Service: http&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/networking.go:198&#xA;STEP: Performing setup for networking test in namespace nettest-6735&#xA;STEP: creating a selector&#xA;STEP: Creating the service pods in kubernetes&#xA;May 20 23:22:35.937: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable&#xA;May 20 23:22:35.968: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 20 23:22:37.972: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 20 23:22:39.971: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 20 23:22:41.972: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 20 23:22:43.972: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 20 23:22:45.971: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 20 23:22:47.972: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:22:49.972: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:22:51.973: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:22:53.972: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:22:55.972: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:22:57.973: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:22:59.971: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:23:01.973: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:23:03.971: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:23:05.971: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:23:07.976: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 20 23:23:09.971: INFO: The status of Pod netserver-0 is Running (Ready = true)&#xA;May 20 23:23:09.977: INFO: The status of Pod netserver-1 is Running (Ready = true)&#xA;STEP: Creating test pods&#xA;May 20 23:23:18.018: INFO: Setting MaxTries for pod polling to 34 for networking test based on endpoint count 2&#xA;STEP: Getting node addresses&#xA;May 20 23:23:18.018: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable&#xA;STEP: Creating the service on top of the pods in kubernetes&#xA;May 20 23:23:18.042: INFO: Service node-port-service in namespace nettest-6735 found.&#xA;May 20 23:23:18.057: INFO: Service session-affinity-service in namespace nettest-6735 found.&#xA;STEP: Waiting for NodePort service to expose endpoint&#xA;May 20 23:23:19.061: INFO: Waiting for amount of service:node-port-service endpoints to be 2&#xA;STEP: Waiting for Session Affinity service to expose endpoint&#xA;May 20 23:23:20.064: INFO: Waiting for amount of service:session-affinity-service endpoints to be 2&#xA;STEP: dialing(http) 10.10.190.207 (node) --&gt; 10.233.40.232:80 (config.clusterIP)&#xA;May 20 23:23:20.067: INFO: Going to poll 10.233.40.232 on port 80 at least 0 times, with a maximum of 34 tries before failing&#xA;May 20 23:23:20.070: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.40.232:80/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:20.070: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:20.478: INFO: Waiting for [netserver-0] endpoints (expected=[netserver-0 netserver-1], actual=[netserver-1])&#xA;May 20 23:23:22.483: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.40.232:80/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:22.483: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:22.584: INFO: Found all 2 expected endpoints: [netserver-0 netserver-1]&#xA;STEP: dialing(http) 10.10.190.207 (node) --&gt; 10.10.190.207:32102 (nodeIP)&#xA;May 20 23:23:22.584: INFO: Going to poll 10.10.190.207 on port 32102 at least 0 times, with a maximum of 34 tries before failing&#xA;May 20 23:23:22.586: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:22.587: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:22.676: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:22.676: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:24.681: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:24.681: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:25.278: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:25.278: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:27.286: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:27.286: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:27.818: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:27.818: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:29.822: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:29.822: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:30.222: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:30.222: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:32.225: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:32.225: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:32.375: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:32.375: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:34.379: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:34.379: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:34.490: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:34.490: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:36.494: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:36.494: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:36.690: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:36.690: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:38.693: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:38.693: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:39.053: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:39.053: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:41.060: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:41.060: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:41.249: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:41.249: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:43.367: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:43.367: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:43.622: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:43.622: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:45.626: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:45.626: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:45.754: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:45.754: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:47.761: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:47.761: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:47.926: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:47.926: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:49.929: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:49.929: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:50.294: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:50.294: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:52.299: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:52.299: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:52.402: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:52.402: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:54.407: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:54.407: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:54.578: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:54.578: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:56.583: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:56.583: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:56.798: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:56.798: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:23:58.803: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:23:58.803: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:23:59.071: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:23:59.071: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:01.076: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:01.076: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:01.204: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:01.204: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:03.207: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:03.208: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:03.363: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:03.363: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:05.367: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:05.367: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:05.497: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:05.497: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:07.502: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:07.502: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:07.585: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:07.585: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:09.591: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:09.591: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:09.677: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:09.677: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:11.682: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:11.682: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:11.863: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:11.863: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:13.868: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:13.868: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:14.036: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:14.036: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:16.040: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:16.040: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:16.264: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:16.264: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:18.268: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:18.268: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:18.356: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:18.356: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:20.360: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:20.360: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:20.475: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:20.475: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:22.478: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:22.478: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:22.612: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:22.612: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:24.614: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:24.614: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:24.710: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:24.710: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:26.713: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:26.713: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:26.813: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:26.813: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:28.817: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:28.817: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:29.111: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:29.111: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:31.118: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:31.118: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:31.215: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:31.215: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:33.218: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:33.218: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:34.327: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:34.327: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:36.330: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\s*$&#39;] Namespace:nettest-6735 PodName:host-test-container-pod ContainerName:agnhost-container Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}&#xA;May 20 23:24:36.330: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 20 23:24:36.516: INFO: Failed to execute &#34;curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName | grep -v &#39;^\\s*$&#39;&#34;: command terminated with exit code 1, stdout: &#34;&#34;, stderr: &#34;&#34;&#xA;May 20 23:24:36.516: INFO: Waiting for [netserver-0 netserver-1] endpoints (expected=[netserver-0 netserver-1], actual=[])&#xA;May 20 23:24:38.517: INFO: &#xA;Output of kubectl describe pod nettest-6735/netserver-0:&#xA;&#xA;May 20 23:24:38.517: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=nettest-6735 describe pod netserver-0 --namespace=nettest-6735&#39;&#xA;May 20 23:24:38.725: INFO: stderr: &#34;&#34;&#xA;May 20 23:24:38.725: INFO: stdout: &#34;Name:         netserver-0\nNamespace:    nettest-6735\nPriority:     0\nNode:         node1/10.10.190.207\nStart Time:   Fri, 20 May 2022 23:22:35 +0000\nLabels:       selector-6cacff6a-9db1-4515-8f9c-9c320e208311=true\nAnnotations:  k8s.v1.cni.cncf.io/network-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.4.179\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;fa:e6:78:d9:cd:4a\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              k8s.v1.cni.cncf.io/networks-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.4.179\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;fa:e6:78:d9:cd:4a\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              kubernetes.io/psp: collectd\nStatus:       Running\nIP:           10.244.4.179\nIPs:\n  IP:  10.244.4.179\nContainers:\n  webserver:\n    Container ID:  docker://1cfba9549c96b9cecc679b748a7fb438538867414c39ba67b6ae40352b2ba4e5\n    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.32\n    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1\n    Ports:         8080/TCP, 8081/UDP\n    Host Ports:    0/TCP, 0/UDP\n    Args:\n      netexec\n      --http-port=8080\n      --udp-port=8081\n    State:          Running\n      Started:      Fri, 20 May 2022 23:22:46 +0000\n    Ready:          True\n    Restart Count:  0\n    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Environment:    &lt;none&gt;\n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-mzb9r (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  kube-api-access-mzb9r:\n    Type:                    Projected (a volume that contains injected data from multiple sources)\n    TokenExpirationSeconds:  3607\n    ConfigMapName:           kube-root-ca.crt\n    ConfigMapOptional:       &lt;nil&gt;\n    DownwardAPI:             true\nQoS Class:                   BestEffort\nNode-Selectors:              kubernetes.io/hostname=node1\nTolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n  Type    Reason     Age   From               Message\n  ----    ------     ----  ----               -------\n  Normal  Scheduled  2m2s  default-scheduler  Successfully assigned nettest-6735/netserver-0 to node1\n  Normal  Pulling    115s  kubelet            Pulling image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.32\&#34;\n  Normal  Pulled     113s  kubelet            Successfully pulled image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.32\&#34; in 1.591861463s\n  Normal  Created    113s  kubelet            Created container webserver\n  Normal  Started    112s  kubelet            Started container webserver\n&#34;&#xA;May 20 23:24:38.725: INFO: Name:         netserver-0&#xA;Namespace:    nettest-6735&#xA;Priority:     0&#xA;Node:         node1/10.10.190.207&#xA;Start Time:   Fri, 20 May 2022 23:22:35 +0000&#xA;Labels:       selector-6cacff6a-9db1-4515-8f9c-9c320e208311=true&#xA;Annotations:  k8s.v1.cni.cncf.io/network-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.4.179&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;fa:e6:78:d9:cd:4a&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              k8s.v1.cni.cncf.io/networks-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.4.179&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;fa:e6:78:d9:cd:4a&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              kubernetes.io/psp: collectd&#xA;Status:       Running&#xA;IP:           10.244.4.179&#xA;IPs:&#xA;  IP:  10.244.4.179&#xA;Containers:&#xA;  webserver:&#xA;    Container ID:  docker://1cfba9549c96b9cecc679b748a7fb438538867414c39ba67b6ae40352b2ba4e5&#xA;    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.32&#xA;    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1&#xA;    Ports:         8080/TCP, 8081/UDP&#xA;    Host Ports:    0/TCP, 0/UDP&#xA;    Args:&#xA;      netexec&#xA;      --http-port=8080&#xA;      --udp-port=8081&#xA;    State:          Running&#xA;      Started:      Fri, 20 May 2022 23:22:46 +0000&#xA;    Ready:          True&#xA;    Restart Count:  0&#xA;    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Environment:    &lt;none&gt;&#xA;    Mounts:&#xA;      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-mzb9r (ro)&#xA;Conditions:&#xA;  Type              Status&#xA;  Initialized       True &#xA;  Ready             True &#xA;  ContainersReady   True &#xA;  PodScheduled      True &#xA;Volumes:&#xA;  kube-api-access-mzb9r:&#xA;    Type:                    Projected (a volume that contains injected data from multiple sources)&#xA;    TokenExpirationSeconds:  3607&#xA;    ConfigMapName:           kube-root-ca.crt&#xA;    ConfigMapOptional:       &lt;nil&gt;&#xA;    DownwardAPI:             true&#xA;QoS Class:                   BestEffort&#xA;Node-Selectors:              kubernetes.io/hostname=node1&#xA;Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s&#xA;                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s&#xA;Events:&#xA;  Type    Reason     Age   From               Message&#xA;  ----    ------     ----  ----               -------&#xA;  Normal  Scheduled  2m2s  default-scheduler  Successfully assigned nettest-6735/netserver-0 to node1&#xA;  Normal  Pulling    115s  kubelet            Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;  Normal  Pulled     113s  kubelet            Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 1.591861463s&#xA;  Normal  Created    113s  kubelet            Created container webserver&#xA;  Normal  Started    112s  kubelet            Started container webserver&#xA;&#xA;May 20 23:24:38.725: INFO: &#xA;Output of kubectl describe pod nettest-6735/netserver-1:&#xA;&#xA;May 20 23:24:38.725: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=nettest-6735 describe pod netserver-1 --namespace=nettest-6735&#39;&#xA;May 20 23:24:38.908: INFO: stderr: &#34;&#34;&#xA;May 20 23:24:38.908: INFO: stdout: &#34;Name:         netserver-1\nNamespace:    nettest-6735\nPriority:     0\nNode:         node2/10.10.190.208\nStart Time:   Fri, 20 May 2022 23:22:35 +0000\nLabels:       selector-6cacff6a-9db1-4515-8f9c-9c320e208311=true\nAnnotations:  k8s.v1.cni.cncf.io/network-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.3.5\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;9e:48:c3:7b:44:ef\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              k8s.v1.cni.cncf.io/networks-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.3.5\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;9e:48:c3:7b:44:ef\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              kubernetes.io/psp: collectd\nStatus:       Running\nIP:           10.244.3.5\nIPs:\n  IP:  10.244.3.5\nContainers:\n  webserver:\n    Container ID:  docker://ef875d73e51e807a8775baace7612016c8097235e36a78645f3bb22733bb7597\n    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.32\n    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1\n    Ports:         8080/TCP, 8081/UDP\n    Host Ports:    0/TCP, 0/UDP\n    Args:\n      netexec\n      --http-port=8080\n      --udp-port=8081\n    State:          Running\n      Started:      Fri, 20 May 2022 23:22:46 +0000\n    Ready:          True\n    Restart Count:  0\n    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Environment:    &lt;none&gt;\n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-gnbd8 (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  kube-api-access-gnbd8:\n    Type:                    Projected (a volume that contains injected data from multiple sources)\n    TokenExpirationSeconds:  3607\n    ConfigMapName:           kube-root-ca.crt\n    ConfigMapOptional:       &lt;nil&gt;\n    DownwardAPI:             true\nQoS Class:                   BestEffort\nNode-Selectors:              kubernetes.io/hostname=node2\nTolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n  Type    Reason     Age   From               Message\n  ----    ------     ----  ----               -------\n  Normal  Scheduled  2m2s  default-scheduler  Successfully assigned nettest-6735/netserver-1 to node2\n  Normal  Pulling    114s  kubelet            Pulling image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.32\&#34;\n  Normal  Pulled     113s  kubelet            Successfully pulled image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.32\&#34; in 1.274279987s\n  Normal  Created    113s  kubelet            Created container webserver\n  Normal  Started    112s  kubelet            Started container webserver\n&#34;&#xA;May 20 23:24:38.908: INFO: Name:         netserver-1&#xA;Namespace:    nettest-6735&#xA;Priority:     0&#xA;Node:         node2/10.10.190.208&#xA;Start Time:   Fri, 20 May 2022 23:22:35 +0000&#xA;Labels:       selector-6cacff6a-9db1-4515-8f9c-9c320e208311=true&#xA;Annotations:  k8s.v1.cni.cncf.io/network-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.3.5&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;9e:48:c3:7b:44:ef&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              k8s.v1.cni.cncf.io/networks-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.3.5&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;9e:48:c3:7b:44:ef&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              kubernetes.io/psp: collectd&#xA;Status:       Running&#xA;IP:           10.244.3.5&#xA;IPs:&#xA;  IP:  10.244.3.5&#xA;Containers:&#xA;  webserver:&#xA;    Container ID:  docker://ef875d73e51e807a8775baace7612016c8097235e36a78645f3bb22733bb7597&#xA;    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.32&#xA;    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1&#xA;    Ports:         8080/TCP, 8081/UDP&#xA;    Host Ports:    0/TCP, 0/UDP&#xA;    Args:&#xA;      netexec&#xA;      --http-port=8080&#xA;      --udp-port=8081&#xA;    State:          Running&#xA;      Started:      Fri, 20 May 2022 23:22:46 +0000&#xA;    Ready:          True&#xA;    Restart Count:  0&#xA;    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Environment:    &lt;none&gt;&#xA;    Mounts:&#xA;      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-gnbd8 (ro)&#xA;Conditions:&#xA;  Type              Status&#xA;  Initialized       True &#xA;  Ready             True &#xA;  ContainersReady   True &#xA;  PodScheduled      True &#xA;Volumes:&#xA;  kube-api-access-gnbd8:&#xA;    Type:                    Projected (a volume that contains injected data from multiple sources)&#xA;    TokenExpirationSeconds:  3607&#xA;    ConfigMapName:           kube-root-ca.crt&#xA;    ConfigMapOptional:       &lt;nil&gt;&#xA;    DownwardAPI:             true&#xA;QoS Class:                   BestEffort&#xA;Node-Selectors:              kubernetes.io/hostname=node2&#xA;Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s&#xA;                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s&#xA;Events:&#xA;  Type    Reason     Age   From               Message&#xA;  ----    ------     ----  ----               -------&#xA;  Normal  Scheduled  2m2s  default-scheduler  Successfully assigned nettest-6735/netserver-1 to node2&#xA;  Normal  Pulling    114s  kubelet            Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;  Normal  Pulled     113s  kubelet            Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 1.274279987s&#xA;  Normal  Created    113s  kubelet            Created container webserver&#xA;  Normal  Started    112s  kubelet            Started container webserver&#xA;&#xA;May 20 23:24:38.909: FAIL: failed dialing endpoint, failed to find expected endpoints, &#xA;tries 34&#xA;Command curl -g -q -s --max-time 15 --connect-timeout 1 http://10.10.190.207:32102/hostName&#xA;retrieved map[]&#xA;expected map[netserver-0:{} netserver-1:{}]&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc001f81200)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x36c&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc001f81200)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:144 +0x2b&#xA;testing.tRunner(0xc001f81200, 0x70f99e8)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1193 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1238 +0x2b3&#xA;[AfterEach] [sig-network] Networking&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186&#xA;STEP: Collecting events from namespace &#34;nettest-6735&#34;.&#xA;STEP: Found 20 events.&#xA;May 20 23:24:38.914: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for host-test-container-pod: { } Scheduled: Successfully assigned nettest-6735/host-test-container-pod to node1&#xA;May 20 23:24:38.914: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-0: { } Scheduled: Successfully assigned nettest-6735/netserver-0 to node1&#xA;May 20 23:24:38.914: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-1: { } Scheduled: Successfully assigned nettest-6735/netserver-1 to node2&#xA;May 20 23:24:38.914: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for test-container-pod: { } Scheduled: Successfully assigned nettest-6735/test-container-pod to node2&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:43 +0000 UTC - event for netserver-0: {kubelet node1} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:44 +0000 UTC - event for netserver-1: {kubelet node2} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:45 +0000 UTC - event for netserver-0: {kubelet node1} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 1.591861463s&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:45 +0000 UTC - event for netserver-0: {kubelet node1} Created: Created container webserver&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:45 +0000 UTC - event for netserver-1: {kubelet node2} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 1.274279987s&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:45 +0000 UTC - event for netserver-1: {kubelet node2} Created: Created container webserver&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:46 +0000 UTC - event for netserver-0: {kubelet node1} Started: Started container webserver&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:22:46 +0000 UTC - event for netserver-1: {kubelet node2} Started: Started container webserver&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:12 +0000 UTC - event for host-test-container-pod: {kubelet node1} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:12 +0000 UTC - event for host-test-container-pod: {kubelet node1} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 386.851225ms&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:13 +0000 UTC - event for host-test-container-pod: {kubelet node1} Created: Created container agnhost-container&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:13 +0000 UTC - event for host-test-container-pod: {kubelet node1} Started: Started container agnhost-container&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:14 +0000 UTC - event for test-container-pod: {kubelet node2} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:14 +0000 UTC - event for test-container-pod: {kubelet node2} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 320.833611ms&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:15 +0000 UTC - event for test-container-pod: {kubelet node2} Created: Created container webserver&#xA;May 20 23:24:38.914: INFO: At 2022-05-20 23:23:15 +0000 UTC - event for test-container-pod: {kubelet node2} Started: Started container webserver&#xA;May 20 23:24:38.917: INFO: POD                      NODE   PHASE    GRACE  CONDITIONS&#xA;May 20 23:24:38.917: INFO: host-test-container-pod  node1  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:10 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:14 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:14 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:09 +0000 UTC  }]&#xA;May 20 23:24:38.917: INFO: netserver-0              node1  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:22:35 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:05 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:05 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:22:35 +0000 UTC  }]&#xA;May 20 23:24:38.917: INFO: netserver-1              node2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:22:35 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:05 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:05 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:22:35 +0000 UTC  }]&#xA;May 20 23:24:38.918: INFO: test-container-pod       node2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:10 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-05-20 23:23:09 +0000 UTC  }]&#xA;May 20 23:24:38.918: INFO: &#xA;May 20 23:24:38.922: INFO: &#xA;Logging node info for node master1&#xA;May 20 23:24:38.925: INFO: Node Info: &amp;Node{ObjectMeta:{master1    b016dcf2-74b7-4456-916a-8ca363b9ccc3 75623 0 2022-05-20 20:01:28 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/control-plane: node-role.kubernetes.io/master: node.kubernetes.io/exclude-from-external-load-balancers:] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.8.2 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubeadm Update v1 2022-05-20 20:01:31 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/control-plane&#34;:{},&#34;f:node-role.kubernetes.io/master&#34;:{},&#34;f:node.kubernetes.io/exclude-from-external-load-balancers&#34;:{}}}}} {kube-controller-manager Update v1 2022-05-20 20:01:45 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {flanneld Update v1 2022-05-20 20:04:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kubelet Update v1 2022-05-20 20:09:00 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}} {nfd-master Update v1 2022-05-20 20:12:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{451201003520 0} {&lt;nil&gt;} 440625980Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234743296 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{406080902496 0} {&lt;nil&gt;} 406080902496 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324579328 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-05-20 20:07:07 +0000 UTC,LastTransitionTime:2022-05-20 20:07:07 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:30 +0000 UTC,LastTransitionTime:2022-05-20 20:01:26 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:30 +0000 UTC,LastTransitionTime:2022-05-20 20:01:26 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:30 +0000 UTC,LastTransitionTime:2022-05-20 20:01:26 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-05-20 23:24:30 +0000 UTC,LastTransitionTime:2022-05-20 20:04:22 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:e9847a94929d4465bdf672fd6e82b77d,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:a01e5bd5-a73c-4ab6-b80a-cab509b05bc6,KernelVersion:3.10.0-1160.66.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.16,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:727687197,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[sirot/netperf-latest@sha256:23929b922bb077cb341450fc1c90ae42e6f75da5d7ce61cd1880b08560a5ff85 sirot/netperf-latest:latest],SizeBytes:282025213,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[k8s.gcr.io/nfd/node-feature-discovery@sha256:74a1cbd82354f148277e20cdce25d57816e355a896bc67f67a0f722164b16945 k8s.gcr.io/nfd/node-feature-discovery:v0.8.2],SizeBytes:108486428,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[quay.io/prometheus-operator/prometheus-operator@sha256:850c86bfeda4389bc9c757a9fd17ca5a090ea6b424968178d4467492cfa13921 quay.io/prometheus-operator/prometheus-operator:v0.44.1],SizeBytes:42617274,},ContainerImage{Names:[k8s.gcr.io/coredns/coredns@sha256:cc8fb77bc2a0541949d1d9320a641b82fd392b0d3d8145469ca4709ae769980e k8s.gcr.io/coredns/coredns:v1.8.0],SizeBytes:42454755,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[localhost:30500/tasextender@sha256:f65735add9b770eec74999948d1a43963106c14a89579d0158e1ec3a1bae070e tasextender:latest localhost:30500/tasextender:0.4],SizeBytes:28910791,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[nginx@sha256:b92d3b942c8b84da889ac3dc6e83bd20ffb8cd2d8298eba92c8b0bf88d52f03e nginx:1.20.1-alpine],SizeBytes:22721538,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 20 23:24:38.926: INFO: &#xA;Logging kubelet events for node master1&#xA;May 20 23:24:38.928: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May 20 23:24:38.943: INFO: kube-multus-ds-amd64-k8cb6 started at 2022-05-20 20:04:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 20 23:24:38.943: INFO: container-registry-65d7c44b96-n94w5 started at 2022-05-20 20:08:47 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May 20 23:24:38.943: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;May 20 23:24:38.943: INFO: prometheus-operator-585ccfb458-bl62n started at 2022-05-20 20:17:13 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:38.943: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;May 20 23:24:38.943: INFO: kube-scheduler-master1 started at 2022-05-20 20:20:27 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-scheduler ready: true, restart count 1&#xA;May 20 23:24:38.943: INFO: kube-apiserver-master1 started at 2022-05-20 20:02:32 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 20 23:24:38.943: INFO: kube-controller-manager-master1 started at 2022-05-20 20:10:37 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-controller-manager ready: true, restart count 3&#xA;May 20 23:24:38.943: INFO: kube-proxy-rgxh2 started at 2022-05-20 20:03:14 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 20 23:24:38.943: INFO: kube-flannel-tzq8g started at 2022-05-20 20:04:10 +0000 UTC (1+1 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 20 23:24:38.943: INFO: node-feature-discovery-controller-cff799f9f-nq7tc started at 2022-05-20 20:11:58 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;May 20 23:24:38.943: INFO: node-exporter-4rvrg started at 2022-05-20 20:17:21 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:38.943: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:38.943: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 20 23:24:39.053: INFO: &#xA;Latency metrics for node master1&#xA;May 20 23:24:39.053: INFO: &#xA;Logging node info for node master2&#xA;May 20 23:24:39.055: INFO: Node Info: &amp;Node{ObjectMeta:{master2    ddc04b08-e43a-4e18-a612-aa3bf7f8411e 75576 0 2022-05-20 20:01:56 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/control-plane: node-role.kubernetes.io/master: node.kubernetes.io/exclude-from-external-load-balancers:] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubeadm Update v1 2022-05-20 20:01:59 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/control-plane&#34;:{},&#34;f:node-role.kubernetes.io/master&#34;:{},&#34;f:node.kubernetes.io/exclude-from-external-load-balancers&#34;:{}}}}} {flanneld Update v1 2022-05-20 20:04:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2022-05-20 20:04:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {kubelet Update v1 2022-05-20 20:14:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{451201003520 0} {&lt;nil&gt;} 440625980Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234743296 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{406080902496 0} {&lt;nil&gt;} 406080902496 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324579328 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-05-20 20:07:03 +0000 UTC,LastTransitionTime:2022-05-20 20:07:03 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:29 +0000 UTC,LastTransitionTime:2022-05-20 20:01:56 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:29 +0000 UTC,LastTransitionTime:2022-05-20 20:01:56 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:29 +0000 UTC,LastTransitionTime:2022-05-20 20:01:56 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-05-20 23:24:29 +0000 UTC,LastTransitionTime:2022-05-20 20:04:15 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:63d829bfe81540169bcb84ee465e884a,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:fc4aead3-0f07-477a-9f91-3902c50ddf48,KernelVersion:3.10.0-1160.66.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.16,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:727687197,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[sirot/netperf-latest@sha256:23929b922bb077cb341450fc1c90ae42e6f75da5d7ce61cd1880b08560a5ff85 sirot/netperf-latest:latest],SizeBytes:282025213,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[k8s.gcr.io/coredns/coredns@sha256:cc8fb77bc2a0541949d1d9320a641b82fd392b0d3d8145469ca4709ae769980e k8s.gcr.io/coredns/coredns:v1.8.0],SizeBytes:42454755,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 20 23:24:39.056: INFO: &#xA;Logging kubelet events for node master2&#xA;May 20 23:24:39.058: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May 20 23:24:39.073: INFO: kube-multus-ds-amd64-97fkc started at 2022-05-20 20:04:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 20 23:24:39.074: INFO: kube-scheduler-master2 started at 2022-05-20 20:02:34 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container kube-scheduler ready: true, restart count 3&#xA;May 20 23:24:39.074: INFO: kube-controller-manager-master2 started at 2022-05-20 20:10:36 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 20 23:24:39.074: INFO: kube-proxy-wfzg2 started at 2022-05-20 20:03:14 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 20 23:24:39.074: INFO: kube-flannel-wj7hl started at 2022-05-20 20:04:10 +0000 UTC (1+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 20 23:24:39.074: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;May 20 23:24:39.074: INFO: coredns-8474476ff8-tjnfw started at 2022-05-20 20:04:46 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May 20 23:24:39.074: INFO: dns-autoscaler-7df78bfcfb-5qj9t started at 2022-05-20 20:04:48 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;May 20 23:24:39.074: INFO: node-exporter-jfg4p started at 2022-05-20 20:17:20 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:39.074: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 20 23:24:39.074: INFO: kube-apiserver-master2 started at 2022-05-20 20:02:34 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.074: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 20 23:24:39.160: INFO: &#xA;Latency metrics for node master2&#xA;May 20 23:24:39.160: INFO: &#xA;Logging node info for node master3&#xA;May 20 23:24:39.163: INFO: Node Info: &amp;Node{ObjectMeta:{master3    f42c1bd6-d828-4857-9180-56c73dcc370f 75784 0 2022-05-20 20:02:05 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/control-plane: node-role.kubernetes.io/master: node.kubernetes.io/exclude-from-external-load-balancers:] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubeadm Update v1 2022-05-20 20:02:08 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/control-plane&#34;:{},&#34;f:node-role.kubernetes.io/master&#34;:{},&#34;f:node.kubernetes.io/exclude-from-external-load-balancers&#34;:{}}}}} {flanneld Update v1 2022-05-20 20:04:14 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2022-05-20 20:04:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {kubelet Update v1 2022-05-20 20:14:42 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{451201003520 0} {&lt;nil&gt;} 440625980Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234743296 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{406080902496 0} {&lt;nil&gt;} 406080902496 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324579328 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-05-20 20:07:09 +0000 UTC,LastTransitionTime:2022-05-20 20:07:09 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:37 +0000 UTC,LastTransitionTime:2022-05-20 20:02:05 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:37 +0000 UTC,LastTransitionTime:2022-05-20 20:02:05 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:37 +0000 UTC,LastTransitionTime:2022-05-20 20:02:05 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-05-20 23:24:37 +0000 UTC,LastTransitionTime:2022-05-20 20:04:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:6a2131d65a6f41c3b857ed7d5f7d9f9f,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:2fa6d1c6-058c-482a-97f3-d7e9e817b36a,KernelVersion:3.10.0-1160.66.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.16,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:727687197,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[k8s.gcr.io/coredns/coredns@sha256:cc8fb77bc2a0541949d1d9320a641b82fd392b0d3d8145469ca4709ae769980e k8s.gcr.io/coredns/coredns:v1.8.0],SizeBytes:42454755,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[aquasec/kube-bench@sha256:3544f6662feb73d36fdba35b17652e2fd73aae45bd4b60e76d7ab928220b3cc6 aquasec/kube-bench:0.3.1],SizeBytes:19301876,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 20 23:24:39.163: INFO: &#xA;Logging kubelet events for node master3&#xA;May 20 23:24:39.165: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May 20 23:24:39.180: INFO: kube-controller-manager-master3 started at 2022-05-20 20:10:36 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Container kube-controller-manager ready: true, restart count 1&#xA;May 20 23:24:39.180: INFO: kube-scheduler-master3 started at 2022-05-20 20:02:33 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Container kube-scheduler ready: true, restart count 2&#xA;May 20 23:24:39.180: INFO: kube-proxy-rsqzq started at 2022-05-20 20:03:14 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 20 23:24:39.180: INFO: kube-flannel-bwb5w started at 2022-05-20 20:04:10 +0000 UTC (1+1 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 20 23:24:39.180: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 20 23:24:39.180: INFO: kube-apiserver-master3 started at 2022-05-20 20:02:05 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 20 23:24:39.180: INFO: kube-multus-ds-amd64-ch8bd started at 2022-05-20 20:04:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 20 23:24:39.180: INFO: coredns-8474476ff8-4szxh started at 2022-05-20 20:04:50 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May 20 23:24:39.180: INFO: node-exporter-zgxkr started at 2022-05-20 20:17:20 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:39.180: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:39.180: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 20 23:24:39.275: INFO: &#xA;Latency metrics for node master3&#xA;May 20 23:24:39.275: INFO: &#xA;Logging node info for node node1&#xA;May 20 23:24:39.278: INFO: Node Info: &amp;Node{ObjectMeta:{node1    65c381dd-b6f5-4e67-a327-7a45366d15af 75732 0 2022-05-20 20:03:10 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.SSE4:true feature.node.kubernetes.io/cpu-cpuid.SSE42:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-cstate.enabled:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.status:active feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.66.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.SSE4,cpu-cpuid.SSE42,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-cstate.enabled,cpu-hardware_multithreading,cpu-pstate.status,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.8.2 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2022-05-20 20:03:10 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2022-05-20 20:03:11 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2022-05-20 20:04:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2022-05-20 20:12:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE4&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE42&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cstate.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.status&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2022-05-20 20:15:36 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2022-05-20 22:31:12 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:example.com/fakecpu&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:scheduling.k8s.io/foo&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}} {e2e.test Update v1 2022-05-20 22:57:29 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:example.com/fakecpu&#34;:{},&#34;f:scheduling.k8s.io/foo&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{451201003520 0} {&lt;nil&gt;} 440625980Ki BinarySI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269608448 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},scheduling.k8s.io/foo: {{3 0} {&lt;nil&gt;} 3 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{406080902496 0} {&lt;nil&gt;} 406080902496 DecimalSI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884608000 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},scheduling.k8s.io/foo: {{3 0} {&lt;nil&gt;} 3 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-05-20 20:07:03 +0000 UTC,LastTransitionTime:2022-05-20 20:07:03 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:35 +0000 UTC,LastTransitionTime:2022-05-20 20:03:10 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:35 +0000 UTC,LastTransitionTime:2022-05-20 20:03:10 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:35 +0000 UTC,LastTransitionTime:2022-05-20 20:03:10 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-05-20 23:24:35 +0000 UTC,LastTransitionTime:2022-05-20 20:04:16 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:f2f0a31e38e446cda6cf4c679d8a2ef5,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:c988afd2-8149-4515-9a6f-832552c2ed2d,KernelVersion:3.10.0-1160.66.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.16,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[opnfv/barometer-collectd@sha256:f30e965aa6195e6ac4ca2410f5a15e3704c92e4afa5208178ca22a7911975d66],SizeBytes:1075575763,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1003977757,},ContainerImage{Names:[localhost:30500/cmk@sha256:1b6fdb10d02a95904d28fbec7317b3044b913b4572405caf5a5b4f305481ce37 cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:727687197,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[aquasec/kube-hunter@sha256:2be6820bc1d7e0f57193a9a27d5a3e16b2fd93c53747b03ce8ca48c6fc323781 aquasec/kube-hunter:0.3.1],SizeBytes:347611549,},ContainerImage{Names:[golang@sha256:db2475a1dbb2149508e5db31d7d77a75e6600d54be645f37681f03f2762169ba golang:alpine3.12],SizeBytes:301186719,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/jessie-dnsutils@sha256:702a992280fb7c3303e84a5801acbb4c9c7fcf48cffe0e9c8be3f0c60f74cf89 k8s.gcr.io/e2e-test-images/jessie-dnsutils:1.4],SizeBytes:253371792,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[grafana/grafana@sha256:ba39bf5131dcc0464134a3ff0e26e8c6380415249fa725e5f619176601255172 grafana/grafana:7.5.4],SizeBytes:203572842,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:b899dbd1b9017b9a379f76ce5b40eead01a62762c4f2057eacef945c3c22d210 quay.io/prometheus/prometheus:v2.22.1],SizeBytes:168344243,},ContainerImage{Names:[nginx@sha256:a05b0cdd4fc1be3b224ba9662ebdf98fe44c09c0c9215b45f84344c12867002e nginx:1.21.1],SizeBytes:133175493,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:716d2f68314c5c4ddd5ecdb45183fcb4ed8019015982c1321571f863989b70b0 k8s.gcr.io/e2e-test-images/httpd:2.4.39-1],SizeBytes:126894770,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50 k8s.gcr.io/e2e-test-images/httpd:2.4.38-1],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nautilus@sha256:1f36a24cfb5e0c3f725d7565a867c2384282fcbeccc77b07b423c9da95763a9a k8s.gcr.io/e2e-test-images/nautilus:1.4],SizeBytes:121748345,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[k8s.gcr.io/nfd/node-feature-discovery@sha256:74a1cbd82354f148277e20cdce25d57816e355a896bc67f67a0f722164b16945 k8s.gcr.io/nfd/node-feature-discovery:v0.8.2],SizeBytes:108486428,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter@sha256:2b09a571757a12c0245f2f1a74db4d1b9386ff901cf57f5ce48a0a682bd0e3af directxman12/k8s-prometheus-adapter:v0.8.2],SizeBytes:68230450,},ContainerImage{Names:[k8s.gcr.io/build-image/debian-iptables@sha256:160595fccf5ad4e41cc0a7acf56027802bf1a2310e704f6505baf0f88746e277 k8s.gcr.io/build-image/debian-iptables:buster-v1.6.7],SizeBytes:60182103,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:bcea5fd975bec7f8eb179f896b3a007090d081bd13d974bdb01eedd94cdd88b1 nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.2],SizeBytes:42676189,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[aquasec/kube-bench@sha256:3544f6662feb73d36fdba35b17652e2fd73aae45bd4b60e76d7ab928220b3cc6 aquasec/kube-bench:0.3.1],SizeBytes:19301876,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nginx@sha256:503b7abb89e57383eba61cc8a9cb0b495ea575c516108f7d972a6ff6e1ab3c9b k8s.gcr.io/e2e-test-images/nginx:1.14-1],SizeBytes:16032814,},ContainerImage{Names:[quay.io/prometheus-operator/prometheus-config-reloader@sha256:4dee0fcf1820355ddd6986c1317b555693776c731315544a99d6cc59a7e34ce9 quay.io/prometheus-operator/prometheus-config-reloader:v0.44.1],SizeBytes:13433274,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:c75ac27b49326926b803b9ed43bf088bc220d22556de1bc5f72d742c91398f69 alpine:3.12],SizeBytes:5581590,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/busybox@sha256:39e1e963e5310e9c313bad51523be012ede7b35bb9316517d19089a010356592 k8s.gcr.io/e2e-test-images/busybox:1.29-1],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 20 23:24:39.280: INFO: &#xA;Logging kubelet events for node node1&#xA;May 20 23:24:39.283: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May 20 23:24:39.318: INFO: kube-multus-ds-amd64-krd6m started at 2022-05-20 20:04:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 20 23:24:39.318: INFO: kubernetes-dashboard-785dcbb76d-6c2f8 started at 2022-05-20 20:04:50 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 2&#xA;May 20 23:24:39.318: INFO: service-proxy-disabled-pmwdl started at 2022-05-20 23:24:26 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container service-proxy-disabled ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-qn9gl started at 2022-05-20 20:13:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: node-exporter-czwvh started at 2022-05-20 20:17:20 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: pod-client started at 2022-05-20 23:23:51 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container pod-client ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: startup-script started at 2022-05-20 23:23:35 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container startup-script ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: netserver-0 started at 2022-05-20 23:23:56 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container webserver ready: false, restart count 0&#xA;May 20 23:24:39.318: INFO: nginx-proxy-node1 started at 2022-05-20 20:06:57 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 20 23:24:39.318: INFO: prometheus-k8s-0 started at 2022-05-20 20:17:30 +0000 UTC (0+4 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container config-reloader ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May 20 23:24:39.318: INFO: e2e-net-exec started at 2022-05-20 23:23:55 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container e2e-net-exec ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: up-down-2-whrhh started at 2022-05-20 23:23:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container up-down-2 ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: iperf2-clients-bqjc8 started at 2022-05-20 23:24:13 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container iperf2-client ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: collectd-875j8 started at 2022-05-20 20:21:17 +0000 UTC (0+3 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: netserver-0 started at 2022-05-20 23:24:16 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: node-feature-discovery-worker-rh55h started at 2022-05-20 20:11:58 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: cmk-init-discover-node1-vkzkd started at 2022-05-20 20:15:33 +0000 UTC (0+3 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 20 23:24:39.318: INFO: service-proxy-disabled-9hsfn started at 2022-05-20 23:24:26 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container service-proxy-disabled ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: netserver-0 started at 2022-05-20 23:22:35 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: kube-flannel-2blt7 started at 2022-05-20 20:04:10 +0000 UTC (1+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 20 23:24:39.318: INFO: &#x9;Container kube-flannel ready: true, restart count 3&#xA;May 20 23:24:39.318: INFO: host-test-container-pod started at 2022-05-20 23:23:10 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container agnhost-container ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: up-down-3-txdwc started at 2022-05-20 23:24:30 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container up-down-3 ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: pod-client started at 2022-05-20 23:23:46 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container pod-client ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: kube-proxy-v8kzq started at 2022-05-20 20:03:14 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 20 23:24:39.318: INFO: cmk-c5x47 started at 2022-05-20 20:16:15 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: up-down-2-6pjb4 started at 2022-05-20 23:23:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.318: INFO: &#x9;Container up-down-2 ready: true, restart count 0&#xA;May 20 23:24:39.318: INFO: service-proxy-disabled-bd8ld started at 2022-05-20 23:24:26 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.319: INFO: &#x9;Container service-proxy-disabled ready: true, restart count 0&#xA;May 20 23:24:39.319: INFO: up-down-3-s8wr5 started at 2022-05-20 23:24:30 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.319: INFO: &#x9;Container up-down-3 ready: true, restart count 0&#xA;May 20 23:24:39.617: INFO: &#xA;Latency metrics for node node1&#xA;May 20 23:24:39.617: INFO: &#xA;Logging node info for node node2&#xA;May 20 23:24:39.620: INFO: Node Info: &amp;Node{ObjectMeta:{node2    a0e0a426-876d-4419-96e4-c6977ef3393c 75711 0 2022-05-20 20:03:09 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.SSE4:true feature.node.kubernetes.io/cpu-cpuid.SSE42:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-cstate.enabled:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.status:active feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.66.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.SSE4,cpu-cpuid.SSE42,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-cstate.enabled,cpu-hardware_multithreading,cpu-pstate.status,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.8.2 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2022-05-20 20:03:09 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2022-05-20 20:03:10 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2022-05-20 20:04:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2022-05-20 20:12:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE4&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE42&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cstate.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.status&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2022-05-20 20:15:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {e2e.test Update v1 2022-05-20 22:31:06 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:scheduling.k8s.io/foo&#34;:{}}}}} {kubelet Update v1 2022-05-20 22:31:12 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:scheduling.k8s.io/foo&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{451201003520 0} {&lt;nil&gt;} 440625980Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269608448 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},scheduling.k8s.io/foo: {{3 0} {&lt;nil&gt;} 3 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{406080902496 0} {&lt;nil&gt;} 406080902496 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884608000 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},scheduling.k8s.io/foo: {{3 0} {&lt;nil&gt;} 3 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-05-20 20:07:03 +0000 UTC,LastTransitionTime:2022-05-20 20:07:03 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:33 +0000 UTC,LastTransitionTime:2022-05-20 20:03:09 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:33 +0000 UTC,LastTransitionTime:2022-05-20 20:03:09 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-05-20 23:24:33 +0000 UTC,LastTransitionTime:2022-05-20 20:03:09 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-05-20 23:24:33 +0000 UTC,LastTransitionTime:2022-05-20 20:07:08 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:a6deb87c5d6d4ca89be50c8f447a0e3c,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:67af2183-25fe-4024-95ea-e80edf7c8695,KernelVersion:3.10.0-1160.66.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.16,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[opnfv/barometer-collectd@sha256:f30e965aa6195e6ac4ca2410f5a15e3704c92e4afa5208178ca22a7911975d66],SizeBytes:1075575763,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:727687197,},ContainerImage{Names:[localhost:30500/cmk@sha256:1b6fdb10d02a95904d28fbec7317b3044b913b4572405caf5a5b4f305481ce37 localhost:30500/cmk:v1.5.1],SizeBytes:727687197,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[sirot/netperf-latest@sha256:23929b922bb077cb341450fc1c90ae42e6f75da5d7ce61cd1880b08560a5ff85 sirot/netperf-latest:latest],SizeBytes:282025213,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[nginx@sha256:a05b0cdd4fc1be3b224ba9662ebdf98fe44c09c0c9215b45f84344c12867002e nginx:1.21.1],SizeBytes:133175493,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:716d2f68314c5c4ddd5ecdb45183fcb4ed8019015982c1321571f863989b70b0 k8s.gcr.io/e2e-test-images/httpd:2.4.39-1],SizeBytes:126894770,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50 k8s.gcr.io/e2e-test-images/httpd:2.4.38-1],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nautilus@sha256:1f36a24cfb5e0c3f725d7565a867c2384282fcbeccc77b07b423c9da95763a9a k8s.gcr.io/e2e-test-images/nautilus:1.4],SizeBytes:121748345,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[k8s.gcr.io/nfd/node-feature-discovery@sha256:74a1cbd82354f148277e20cdce25d57816e355a896bc67f67a0f722164b16945 k8s.gcr.io/nfd/node-feature-discovery:v0.8.2],SizeBytes:108486428,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/sample-apiserver@sha256:e7fddbaac4c3451da2365ab90bad149d32f11409738034e41e0f460927f7c276 k8s.gcr.io/e2e-test-images/sample-apiserver:1.17.4],SizeBytes:58172101,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/regression-issue-74839@sha256:b4f1d8d61bdad84bd50442d161d5460e4019d53e989b64220fdbc62fc87d76bf k8s.gcr.io/e2e-test-images/regression-issue-74839:1.2],SizeBytes:44576952,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:bcea5fd975bec7f8eb179f896b3a007090d081bd13d974bdb01eedd94cdd88b1 localhost:30500/sriov-device-plugin:v3.3.2],SizeBytes:42676189,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[localhost:30500/tasextender@sha256:f65735add9b770eec74999948d1a43963106c14a89579d0158e1ec3a1bae070e localhost:30500/tasextender:0.4],SizeBytes:28910791,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nginx@sha256:503b7abb89e57383eba61cc8a9cb0b495ea575c516108f7d972a6ff6e1ab3c9b k8s.gcr.io/e2e-test-images/nginx:1.14-1],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nonewprivs@sha256:8ac1264691820febacf3aea5d152cbde6d10685731ec14966a9401c6f47a68ac k8s.gcr.io/e2e-test-images/nonewprivs:1.3],SizeBytes:7107254,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/busybox@sha256:39e1e963e5310e9c313bad51523be012ede7b35bb9316517d19089a010356592 k8s.gcr.io/e2e-test-images/busybox:1.29-1],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 20 23:24:39.621: INFO: &#xA;Logging kubelet events for node node2&#xA;May 20 23:24:39.623: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May 20 23:24:39.640: INFO: cmk-webhook-6c9d5f8578-5kbbc started at 2022-05-20 20:16:16 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.640: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May 20 23:24:39.640: INFO: up-down-3-qrm62 started at 2022-05-20 23:24:30 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.640: INFO: &#x9;Container up-down-3 ready: true, restart count 0&#xA;May 20 23:24:39.640: INFO: service-proxy-toggled-zg2kg started at 2022-05-20 23:24:35 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.640: INFO: &#x9;Container service-proxy-toggled ready: false, restart count 0&#xA;May 20 23:24:39.640: INFO: netserver-1 started at 2022-05-20 23:23:56 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.640: INFO: &#x9;Container webserver ready: false, restart count 0&#xA;May 20 23:24:39.640: INFO: kube-multus-ds-amd64-p22zp started at 2022-05-20 20:04:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.640: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 20 23:24:39.640: INFO: kubernetes-metrics-scraper-5558854cb-66r9g started at 2022-05-20 20:04:50 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.640: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May 20 23:24:39.641: INFO: tas-telemetry-aware-scheduling-84ff454dfb-ddzzd started at 2022-05-20 20:20:26 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: boom-server started at 2022-05-20 23:23:27 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container boom-server ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: cmk-init-discover-node2-b7gw4 started at 2022-05-20 20:15:53 +0000 UTC (0+3 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 20 23:24:39.641: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 20 23:24:39.641: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 20 23:24:39.641: INFO: collectd-h4pzk started at 2022-05-20 20:21:17 +0000 UTC (0+3 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: up-down-2-b7fkw started at 2022-05-20 23:23:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container up-down-2 ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: nodeport-update-service-skqd5 started at 2022-05-20 23:24:13 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container nodeport-update-service ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: nginx-proxy-node2 started at 2022-05-20 20:03:09 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 20 23:24:39.641: INFO: kube-proxy-rg2fp started at 2022-05-20 20:03:14 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 20 23:24:39.641: INFO: kube-flannel-jpmpd started at 2022-05-20 20:04:10 +0000 UTC (1+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Init container install-cni ready: true, restart count 1&#xA;May 20 23:24:39.641: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 20 23:24:39.641: INFO: verify-service-up-host-exec-pod started at 2022-05-20 23:24:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container agnhost-container ready: false, restart count 0&#xA;May 20 23:24:39.641: INFO: iperf2-server-deployment-59979d877-bmv7x started at 2022-05-20 23:24:09 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container iperf2-server ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: netserver-1 started at 2022-05-20 23:22:35 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: pod-server-1 started at 2022-05-20 23:23:54 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container agnhost-container ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: iperf2-clients-ql5nt started at 2022-05-20 23:24:13 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container iperf2-client ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: netserver-1 started at 2022-05-20 23:24:16 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: service-proxy-toggled-x4qjm started at 2022-05-20 23:24:35 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container service-proxy-toggled ready: false, restart count 0&#xA;May 20 23:24:39.641: INFO: node-feature-discovery-worker-nphk9 started at 2022-05-20 20:11:58 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: service-proxy-toggled-bch5g started at 2022-05-20 23:24:35 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container service-proxy-toggled ready: false, restart count 0&#xA;May 20 23:24:39.641: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-wl7nk started at 2022-05-20 20:13:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: cmk-9hxtl started at 2022-05-20 20:16:16 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: node-exporter-vm24n started at 2022-05-20 20:17:20 +0000 UTC (0+2 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: pod-server-2 started at 2022-05-20 23:24:13 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container agnhost-container ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: test-container-pod started at 2022-05-20 23:23:10 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: test-container-pod started at 2022-05-20 23:24:18 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container webserver ready: false, restart count 0&#xA;May 20 23:24:39.641: INFO: nodeport-update-service-lrn9m started at 2022-05-20 23:24:13 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container nodeport-update-service ready: true, restart count 0&#xA;May 20 23:24:39.641: INFO: execpodpnrdf started at 2022-05-20 23:24:26 +0000 UTC (0+1 container statuses recorded)&#xA;May 20 23:24:39.641: INFO: &#x9;Container agnhost-container ready: true, restart count 0&#xA;May 20 23:24:40.428: INFO: &#xA;Latency metrics for node node2&#xA;May 20 23:24:40.428: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;nettest-6735&#34; for this suite.&#xA;</system-out>
      </testcase>
  </testsuite>