<?xml version="1.0" encoding="UTF-8"?>
  <testsuite name="Kubernetes e2e suite" tests="2" failures="1" errors="0" time="177.433">
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] LoadBalancers should reconcile LB health check interval [Slow][Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] PreStop graceful pod terminated should wait until preStop hook completes the process" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI FSGroupPolicy [LinuxOnly] should not modify fsGroup if fsGroupPolicy=None" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap should be immutable if `immutable` field is set [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Variable Expansion should allow substituting values in a volume subpath [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Dynamic Provisioning Invalid AWS KMS key should report an error and create no PV" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with different fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI Volume expansion should not expand volume if resizingOnDriver=off, resizingOnSC=on" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume FStype [Feature:vsphere] verify fstype - default value should be ext4" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should work after restarting kube-proxy [Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] Should be able to scale a node group down to 0[Feature:ClusterSizeAutoscalingScaleDown]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] NFSPersistentVolumes[Disruptive][Flaky] when kubelet restarts Should test that a volume mounted to a pod that is force deleted while the kubelet is down unmounts when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: dir-link-bindmounted] One pod requesting one prebound PVC should be able to mount volume and write from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] ConfigMap should update ConfigMap successfully" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  StatefulSet with pod affinity [Slow] should use volumes spread across nodes when pod management is parallel and pod has anti-affinity" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] NFSPersistentVolumes[Disruptive][Flaky] when kubelet restarts Should test that a file written to the mount before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType File [Slow] Should fail on mounting file &#39;afile&#39; when HostPathType is HostPathDirectory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStack] [LinuxOnly] Granular Checks: Services Secondary IP Family should be able to handle large requests: udp" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] DNS horizontal autoscaling [Serial] [Slow] kube-dns-autoscaler should scale kube-dns pods when cluster size changed" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] ReplicationController should test the lifecycle of a ReplicationController [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] NodeLease when the NodeLease feature is enabled the kubelet should create and update a lease in the kube-node-lease namespace" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Loadbalancing: L7 GCE [Slow] [Feature:kubemci] should remove clusters as expected" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PVC and a pre-bound PV: test write access" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Security Context should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce updated policy [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] [Feature:IPv6DualStack] [LinuxOnly] should create service with ipv6,v4 cluster ip [Feature:IPv6DualStack]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should scale up when non expendable pod is created [Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  Pod with node different from PV&#39;s NodeAffinity should fail scheduling due to different NodeAffinity" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Update Demo should create and stop a replication controller  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Provisioning On Clustered Datastore [Feature:vsphere] verify dynamic provision with default parameter on clustered datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cloud-provider-gcp] HA-master [Feature:HAMaster] survive addition/removal replicas multizone workers [Serial][Disruptive]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] [Serial] Volume metrics should create prometheus metrics for volume provisioning errors [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] EndpointSlice should have Endpoints and EndpointSlices pointing to API Server [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume FStype [Feature:vsphere] verify fstype - ext3 formatted volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl client-side validation should create/apply a valid CR for CRD with validation schema" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Secrets Should fail non-optional pod creation due to secret object does not exist [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS should support configurable pod DNS nameservers [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota [Feature:ScopeSelectors] should verify ResourceQuota with best effort scope using scope-selectors." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] ReplicationController should adopt matching pods on creation [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSI workload information using mock driver should not be passed when podInfoOnMount=false" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ResourceQuota [Feature:CrossNamespacePodAffinity] [alpha] should verify ResourceQuota with cross namespace pod affinity scope using scope-selectors." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: blockfswithoutformat] One pod requesting one prebound PVC should be able to mount volume and write from pod1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Secrets Should fail non-optional pod creation due to the key in the secret object does not exist [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Probing container should be ready immediately after startupProbe succeeds" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] SCTP [Feature:SCTP] [LinuxOnly] should create a ClusterIP Service with SCTP ports" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cloud-provider-gcp] Downgrade [Feature:Downgrade] cluster downgrade should maintain a functioning cluster [Feature:ClusterDowngrade]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup skips ownership changes to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] ReplicaSet Replicaset should have a working scale subresource [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  Pod with node different from PV&#39;s NodeAffinity should fail scheduling due to different NodeSelector" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works  [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] ServerSideApply should work for subresources" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client should deny ingress from pods on other namespaces [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS configMap nameserver Change stubDomain should be able to change stubDomain configuration [Slow][Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap Should fail non-optional pod creation due to configMap object does not exist [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] Device Plugin should be able to create a functioning device plugin for Windows" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume storage capacity exhausted, immediate binding" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Discovery Custom resource should have storage version hash" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-windows] [Feature:Windows] Windows volume mounts  check volume mount permissions container should have readOnly permissions on emptyDir" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Probing container should be restarted with a local redirect http liveness probe" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes NFS with multiple PVs and PVCs all in same ns should create 3 PVs and 3 PVCs: test write access" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-api-machinery] Server request timeout should return HTTP status code 400 if the user specifies an invalid timeout in the request URL" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Provisioning On Clustered Datastore [Feature:vsphere] verify static provisioning on clustered datastore" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Firewall rule control plane should not expose well-known ports" classname="Kubernetes e2e suite" time="0.032383978">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should preserve source pod IP for traffic thru service cluster IP [LinuxOnly]" classname="Kubernetes e2e suite" time="23.171691438"></testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] ReplicaSet Should scale from 5 pods to 3 pods and from 3 to 1" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaler scalability [Slow] should scale down empty nodes [Feature:ClusterAutoscalerScalability3]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Deployment should run the lifecycle of a Deployment [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] CronJob should schedule multiple jobs concurrently [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs should create a non-pre-bound PV and PVC: test write access " classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Character Device [Slow] Should be able to mount character device &#39;achardev&#39; successfully when HostPathType is HostPathCharDev" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: block] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PV Protection Verify &#34;immediate&#34; deletion of a PV that is not bound to a PVC" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] GenericPersistentVolume[Disruptive] When kubelet restarts Should test that a file written to the mount before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pod requesting volume is pending [Feature:ClusterSizeAutoscalingScaleUp]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects a client request should support a client that connects, sends DATA, and disconnects" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] [Feature:NodeAuthorizer] Getting a non-existent secret should exit with the Forbidden error, not a NotFound error" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by ordering clean reboot and ensure they function upon restart" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Directory [Slow] Should fail on mounting non-existent directory &#39;does-not-exist-dir&#39; when HostPathType is HostPathDirectory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Variable Expansion should allow substituting values in a container&#39;s command [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client should enforce ingress policy allowing any port traffic to a server on a specific protocol [Feature:NetworkPolicy] [Feature:UDP]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] NetworkPolicy [LinuxOnly] NetworkPolicy between server and client should enforce policy to allow traffic only from a pod in a different namespace based on PodSelector and NamespaceSelector [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] gpu Upgrade [Feature:GPUUpgrade] master upgrade should NOT disrupt gpu pod [Feature:GPUMasterUpgrade]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Netpol [Feature:SCTPConnectivity][LinuxOnly][Disruptive] NetworkPolicy between server and client using SCTP should enforce policy based on Ports [Feature:NetworkPolicy]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  [Volume type: gce-localssd-scsi-fs] [Serial] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Security Context should support seccomp runtime/default [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ntfs)][Feature:Windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed in first pod, new pod with same fsgroup applied to the volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should adopt matching orphans and release non-matching pods" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable[Feature:VolumeSnapshotDataSource] volume snapshot controller  should check snapshot fields, check restore correctly works after modifying source data, check deletion" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-apps] DisruptionController evictions: maxUnavailable deny evictions, integer =&gt; should not allow an eviction [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand Verify if offline PVC expansion works" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] volume on default medium should have the correct mode using FSGroup" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (ntfs)][Feature:Windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PVC Protection Verify that scheduling of a pod that uses PVC that is being deleted fails and the pod becomes Unschedulable" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-autoscaling] Cluster size autoscaler scalability [Slow] should scale up at all [Feature:ClusterAutoscalerScalability1]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] PersistentVolumes-local  Local volume that cannot be mounted [Slow] should fail due to non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Probing container should *not* be restarted by liveness probe because startup probe delays it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-node] Security Context should support seccomp unconfined on the pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Pre-provisioned PV (xfs)][Slow] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] files with FSGroup ownership should support (root,0644,tmpfs)" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing single file [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-cli] Kubectl client Simple pod should return command exit codes" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (ntfs)][Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] SCTP [Feature:SCTP] [LinuxOnly] should allow creating a basic SCTP service with pod and endpoints" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be able to update service type to NodePort listening on same port number but different protocols" classname="Kubernetes e2e suite" time="152.732106511">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:1211&#xA;Nov 22 02:29:42.372: Unexpected error:&#xA;    &lt;*errors.errorString | 0xc003ab4290&gt;: {&#xA;        s: &#34;service is not reachable within 2m0s timeout on endpoint 10.10.190.207:32493 over TCP protocol&#34;,&#xA;    }&#xA;    service is not reachable within 2m0s timeout on endpoint 10.10.190.207:32493 over TCP protocol&#xA;occurred&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:1245</failure>
          <system-out>[BeforeEach] [sig-network] Services&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185&#xA;STEP: Creating a kubernetes client&#xA;Nov 22 02:27:10.601: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename services&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[BeforeEach] [sig-network] Services&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:746&#xA;[It] should be able to update service type to NodePort listening on same port number but different protocols&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:1211&#xA;STEP: creating a TCP service nodeport-update-service with type=ClusterIP in namespace services-224&#xA;Nov 22 02:27:10.628: INFO: Service Port TCP: 80&#xA;STEP: changing the TCP service to type=NodePort&#xA;STEP: creating replication controller nodeport-update-service in namespace services-224&#xA;I1122 02:27:10.640864      28 runners.go:190] Created replication controller with name: nodeport-update-service, namespace: services-224, replica count: 2&#xA;I1122 02:27:13.694042      28 runners.go:190] nodeport-update-service Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady &#xA;I1122 02:27:16.694290      28 runners.go:190] nodeport-update-service Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady &#xA;I1122 02:27:19.695339      28 runners.go:190] nodeport-update-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady &#xA;Nov 22 02:27:19.695: INFO: Creating new exec pod&#xA;Nov 22 02:27:38.721: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 nodeport-update-service 80&#39;&#xA;Nov 22 02:27:38.995: INFO: stderr: &#34;+ echo hostName\n+ nc -v -t -w 2 nodeport-update-service 80\nConnection to nodeport-update-service 80 port [tcp/http] succeeded!\n&#34;&#xA;Nov 22 02:27:38.995: INFO: stdout: &#34;nodeport-update-service-246qt&#34;&#xA;Nov 22 02:27:38.995: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.15.147 80&#39;&#xA;Nov 22 02:27:39.240: INFO: stderr: &#34;+ echo hostName\n+ nc -v -t -w 2 10.233.15.147 80\nConnection to 10.233.15.147 80 port [tcp/http] succeeded!\n&#34;&#xA;Nov 22 02:27:39.240: INFO: stdout: &#34;&#34;&#xA;Nov 22 02:27:40.241: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.15.147 80&#39;&#xA;Nov 22 02:27:40.513: INFO: stderr: &#34;+ echo hostName\n+ nc -v -t -w 2 10.233.15.147 80\nConnection to 10.233.15.147 80 port [tcp/http] succeeded!\n&#34;&#xA;Nov 22 02:27:40.513: INFO: stdout: &#34;&#34;&#xA;Nov 22 02:27:41.241: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.15.147 80&#39;&#xA;Nov 22 02:27:41.622: INFO: stderr: &#34;+ echo hostName\n+ nc -v -t -w 2 10.233.15.147 80\nConnection to 10.233.15.147 80 port [tcp/http] succeeded!\n&#34;&#xA;Nov 22 02:27:41.622: INFO: stdout: &#34;nodeport-update-service-jgrp8&#34;&#xA;Nov 22 02:27:41.622: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:41.864: INFO: rc: 1&#xA;Nov 22 02:27:41.864: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:42.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:43.377: INFO: rc: 1&#xA;Nov 22 02:27:43.377: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:43.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:44.302: INFO: rc: 1&#xA;Nov 22 02:27:44.302: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:44.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:45.117: INFO: rc: 1&#xA;Nov 22 02:27:45.117: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:45.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:46.136: INFO: rc: 1&#xA;Nov 22 02:27:46.136: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:46.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:47.352: INFO: rc: 1&#xA;Nov 22 02:27:47.352: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:47.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:48.129: INFO: rc: 1&#xA;Nov 22 02:27:48.129: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:48.864: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:49.106: INFO: rc: 1&#xA;Nov 22 02:27:49.106: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:49.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:50.212: INFO: rc: 1&#xA;Nov 22 02:27:50.212: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:50.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:51.129: INFO: rc: 1&#xA;Nov 22 02:27:51.129: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:51.867: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:52.118: INFO: rc: 1&#xA;Nov 22 02:27:52.118: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:52.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:53.254: INFO: rc: 1&#xA;Nov 22 02:27:53.254: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:53.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:54.385: INFO: rc: 1&#xA;Nov 22 02:27:54.385: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:54.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:55.154: INFO: rc: 1&#xA;Nov 22 02:27:55.154: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:55.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:56.161: INFO: rc: 1&#xA;Nov 22 02:27:56.161: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:56.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:57.116: INFO: rc: 1&#xA;Nov 22 02:27:57.117: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:57.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:58.120: INFO: rc: 1&#xA;Nov 22 02:27:58.120: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:58.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:27:59.208: INFO: rc: 1&#xA;Nov 22 02:27:59.208: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:27:59.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:00.157: INFO: rc: 1&#xA;Nov 22 02:28:00.157: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;+ echo hostName&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:00.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:01.203: INFO: rc: 1&#xA;Nov 22 02:28:01.203: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:01.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:02.294: INFO: rc: 1&#xA;Nov 22 02:28:02.294: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:02.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:03.113: INFO: rc: 1&#xA;Nov 22 02:28:03.113: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:03.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:04.150: INFO: rc: 1&#xA;Nov 22 02:28:04.150: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:04.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:05.358: INFO: rc: 1&#xA;Nov 22 02:28:05.358: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:05.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:06.121: INFO: rc: 1&#xA;Nov 22 02:28:06.121: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:06.867: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:07.280: INFO: rc: 1&#xA;Nov 22 02:28:07.280: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:07.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:08.102: INFO: rc: 1&#xA;Nov 22 02:28:08.103: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:08.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:09.111: INFO: rc: 1&#xA;Nov 22 02:28:09.111: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:09.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:10.126: INFO: rc: 1&#xA;Nov 22 02:28:10.126: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:10.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:11.211: INFO: rc: 1&#xA;Nov 22 02:28:11.211: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:11.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:12.180: INFO: rc: 1&#xA;Nov 22 02:28:12.180: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:12.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:13.208: INFO: rc: 1&#xA;Nov 22 02:28:13.208: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:13.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:14.126: INFO: rc: 1&#xA;Nov 22 02:28:14.126: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:14.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:15.209: INFO: rc: 1&#xA;Nov 22 02:28:15.209: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:15.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:16.128: INFO: rc: 1&#xA;Nov 22 02:28:16.128: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:16.867: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:17.181: INFO: rc: 1&#xA;Nov 22 02:28:17.181: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:17.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:18.129: INFO: rc: 1&#xA;Nov 22 02:28:18.129: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:18.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:19.149: INFO: rc: 1&#xA;Nov 22 02:28:19.149: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:19.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:20.122: INFO: rc: 1&#xA;Nov 22 02:28:20.122: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:20.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:21.471: INFO: rc: 1&#xA;Nov 22 02:28:21.471: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:21.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:22.123: INFO: rc: 1&#xA;Nov 22 02:28:22.123: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName+ &#xA;nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:22.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:23.109: INFO: rc: 1&#xA;Nov 22 02:28:23.109: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:23.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:24.513: INFO: rc: 1&#xA;Nov 22 02:28:24.513: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:24.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:25.243: INFO: rc: 1&#xA;Nov 22 02:28:25.243: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:25.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:26.462: INFO: rc: 1&#xA;Nov 22 02:28:26.462: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:26.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:27.431: INFO: rc: 1&#xA;Nov 22 02:28:27.431: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:27.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:28.575: INFO: rc: 1&#xA;Nov 22 02:28:28.575: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:28.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:29.352: INFO: rc: 1&#xA;Nov 22 02:28:29.352: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:29.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:30.385: INFO: rc: 1&#xA;Nov 22 02:28:30.386: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:30.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:31.684: INFO: rc: 1&#xA;Nov 22 02:28:31.684: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:31.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:32.310: INFO: rc: 1&#xA;Nov 22 02:28:32.310: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:32.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:33.308: INFO: rc: 1&#xA;Nov 22 02:28:33.308: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:33.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:34.686: INFO: rc: 1&#xA;Nov 22 02:28:34.686: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:34.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:35.232: INFO: rc: 1&#xA;Nov 22 02:28:35.232: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:35.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:36.140: INFO: rc: 1&#xA;Nov 22 02:28:36.140: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:36.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:37.298: INFO: rc: 1&#xA;Nov 22 02:28:37.298: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:37.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:38.154: INFO: rc: 1&#xA;Nov 22 02:28:38.154: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:38.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:39.279: INFO: rc: 1&#xA;Nov 22 02:28:39.279: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:39.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:40.100: INFO: rc: 1&#xA;Nov 22 02:28:40.100: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:40.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:41.131: INFO: rc: 1&#xA;Nov 22 02:28:41.131: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:41.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:42.107: INFO: rc: 1&#xA;Nov 22 02:28:42.107: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:42.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:43.355: INFO: rc: 1&#xA;Nov 22 02:28:43.355: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:43.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:44.138: INFO: rc: 1&#xA;Nov 22 02:28:44.139: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:44.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:45.096: INFO: rc: 1&#xA;Nov 22 02:28:45.096: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:45.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:46.238: INFO: rc: 1&#xA;Nov 22 02:28:46.238: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:46.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:47.198: INFO: rc: 1&#xA;Nov 22 02:28:47.198: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:47.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:48.097: INFO: rc: 1&#xA;Nov 22 02:28:48.098: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;+ echo hostName&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:48.867: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:49.184: INFO: rc: 1&#xA;Nov 22 02:28:49.184: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:49.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:50.438: INFO: rc: 1&#xA;Nov 22 02:28:50.438: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:50.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:51.447: INFO: rc: 1&#xA;Nov 22 02:28:51.447: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:51.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:52.441: INFO: rc: 1&#xA;Nov 22 02:28:52.441: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ + ncecho -v -t -w 2 hostName 10.10.190.207&#xA; 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:52.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:53.119: INFO: rc: 1&#xA;Nov 22 02:28:53.119: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:53.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:54.192: INFO: rc: 1&#xA;Nov 22 02:28:54.192: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:54.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:55.105: INFO: rc: 1&#xA;Nov 22 02:28:55.105: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:55.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:56.102: INFO: rc: 1&#xA;Nov 22 02:28:56.102: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:56.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:57.092: INFO: rc: 1&#xA;Nov 22 02:28:57.092: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:57.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:58.106: INFO: rc: 1&#xA;Nov 22 02:28:58.106: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:58.867: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:28:59.123: INFO: rc: 1&#xA;Nov 22 02:28:59.123: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:28:59.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:00.094: INFO: rc: 1&#xA;Nov 22 02:29:00.094: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:00.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:01.096: INFO: rc: 1&#xA;Nov 22 02:29:01.096: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:01.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:02.241: INFO: rc: 1&#xA;Nov 22 02:29:02.241: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:02.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:03.157: INFO: rc: 1&#xA;Nov 22 02:29:03.157: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:03.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:04.197: INFO: rc: 1&#xA;Nov 22 02:29:04.197: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:04.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:05.130: INFO: rc: 1&#xA;Nov 22 02:29:05.130: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:05.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:06.400: INFO: rc: 1&#xA;Nov 22 02:29:06.400: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ + nc -vecho -t -w 2 10.10.190.207 32493 hostName&#xA;&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:06.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:07.284: INFO: rc: 1&#xA;Nov 22 02:29:07.284: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:07.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:08.182: INFO: rc: 1&#xA;Nov 22 02:29:08.182: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:08.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:09.094: INFO: rc: 1&#xA;Nov 22 02:29:09.094: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:09.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:10.124: INFO: rc: 1&#xA;Nov 22 02:29:10.124: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:10.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:11.101: INFO: rc: 1&#xA;Nov 22 02:29:11.101: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:11.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:12.089: INFO: rc: 1&#xA;Nov 22 02:29:12.089: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:12.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:13.121: INFO: rc: 1&#xA;Nov 22 02:29:13.121: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;+ echo hostName&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:13.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:14.110: INFO: rc: 1&#xA;Nov 22 02:29:14.110: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:14.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:15.342: INFO: rc: 1&#xA;Nov 22 02:29:15.343: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:15.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:16.124: INFO: rc: 1&#xA;Nov 22 02:29:16.124: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:16.867: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:17.161: INFO: rc: 1&#xA;Nov 22 02:29:17.162: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:17.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:18.109: INFO: rc: 1&#xA;Nov 22 02:29:18.109: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:18.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:19.100: INFO: rc: 1&#xA;Nov 22 02:29:19.100: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:19.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:20.109: INFO: rc: 1&#xA;Nov 22 02:29:20.109: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:20.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:21.103: INFO: rc: 1&#xA;Nov 22 02:29:21.103: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:21.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:22.142: INFO: rc: 1&#xA;Nov 22 02:29:22.142: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:22.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:23.124: INFO: rc: 1&#xA;Nov 22 02:29:23.124: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:23.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:24.114: INFO: rc: 1&#xA;Nov 22 02:29:24.114: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:24.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:25.122: INFO: rc: 1&#xA;Nov 22 02:29:25.122: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:25.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:26.105: INFO: rc: 1&#xA;Nov 22 02:29:26.105: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:26.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:27.114: INFO: rc: 1&#xA;Nov 22 02:29:27.114: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:27.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:28.135: INFO: rc: 1&#xA;Nov 22 02:29:28.135: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:28.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:29.139: INFO: rc: 1&#xA;Nov 22 02:29:29.139: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:29.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:30.092: INFO: rc: 1&#xA;Nov 22 02:29:30.092: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:30.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:31.111: INFO: rc: 1&#xA;Nov 22 02:29:31.111: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:31.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:32.106: INFO: rc: 1&#xA;Nov 22 02:29:32.106: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:32.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:33.102: INFO: rc: 1&#xA;Nov 22 02:29:33.102: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:33.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:34.102: INFO: rc: 1&#xA;Nov 22 02:29:34.102: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:34.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:35.090: INFO: rc: 1&#xA;Nov 22 02:29:35.091: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:35.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:36.125: INFO: rc: 1&#xA;Nov 22 02:29:36.125: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:36.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:37.268: INFO: rc: 1&#xA;Nov 22 02:29:37.268: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:37.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:38.105: INFO: rc: 1&#xA;Nov 22 02:29:38.105: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:38.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:39.133: INFO: rc: 1&#xA;Nov 22 02:29:39.133: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:39.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:40.093: INFO: rc: 1&#xA;Nov 22 02:29:40.093: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:40.865: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:41.097: INFO: rc: 1&#xA;Nov 22 02:29:41.097: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:41.866: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:42.125: INFO: rc: 1&#xA;Nov 22 02:29:42.125: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:42.125: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493&#39;&#xA;Nov 22 02:29:42.371: INFO: rc: 1&#xA;Nov 22 02:29:42.371: INFO: Service reachability failing with error: error running /usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=services-224 exec execpod8f26v -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.10.190.207 32493:&#xA;Command stdout:&#xA;&#xA;stderr:&#xA;+ echo hostName&#xA;+ nc -v -t -w 2 10.10.190.207 32493&#xA;nc: connect to 10.10.190.207 port 32493 (tcp) failed: Connection refused&#xA;command terminated with exit code 1&#xA;&#xA;error:&#xA;exit status 1&#xA;Retrying...&#xA;Nov 22 02:29:42.372: FAIL: Unexpected error:&#xA;    &lt;*errors.errorString | 0xc003ab4290&gt;: {&#xA;        s: &#34;service is not reachable within 2m0s timeout on endpoint 10.10.190.207:32493 over TCP protocol&#34;,&#xA;    }&#xA;    service is not reachable within 2m0s timeout on endpoint 10.10.190.207:32493 over TCP protocol&#xA;occurred&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/network.glob..func24.13()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:1245 +0x431&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc000a49980)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x36c&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc000a49980)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:144 +0x2b&#xA;testing.tRunner(0xc000a49980, 0x70e7b58)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1193 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1238 +0x2b3&#xA;Nov 22 02:29:42.373: INFO: Cleaning up the updating NodePorts test service&#xA;[AfterEach] [sig-network] Services&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186&#xA;STEP: Collecting events from namespace &#34;services-224&#34;.&#xA;STEP: Found 17 events.&#xA;Nov 22 02:29:42.397: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for execpod8f26v: { } Scheduled: Successfully assigned services-224/execpod8f26v to node2&#xA;Nov 22 02:29:42.397: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for nodeport-update-service-246qt: { } Scheduled: Successfully assigned services-224/nodeport-update-service-246qt to node2&#xA;Nov 22 02:29:42.397: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for nodeport-update-service-jgrp8: { } Scheduled: Successfully assigned services-224/nodeport-update-service-jgrp8 to node2&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:10 +0000 UTC - event for nodeport-update-service: {replication-controller } SuccessfulCreate: Created pod: nodeport-update-service-246qt&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:10 +0000 UTC - event for nodeport-update-service: {replication-controller } SuccessfulCreate: Created pod: nodeport-update-service-jgrp8&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:14 +0000 UTC - event for nodeport-update-service-246qt: {kubelet node2} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:14 +0000 UTC - event for nodeport-update-service-246qt: {kubelet node2} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 275.414117ms&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:14 +0000 UTC - event for nodeport-update-service-246qt: {kubelet node2} Created: Created container nodeport-update-service&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:14 +0000 UTC - event for nodeport-update-service-jgrp8: {kubelet node2} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:15 +0000 UTC - event for nodeport-update-service-246qt: {kubelet node2} Started: Started container nodeport-update-service&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:15 +0000 UTC - event for nodeport-update-service-jgrp8: {kubelet node2} Started: Started container nodeport-update-service&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:15 +0000 UTC - event for nodeport-update-service-jgrp8: {kubelet node2} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 550.755183ms&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:15 +0000 UTC - event for nodeport-update-service-jgrp8: {kubelet node2} Created: Created container nodeport-update-service&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:24 +0000 UTC - event for execpod8f26v: {kubelet node2} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34;&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:24 +0000 UTC - event for execpod8f26v: {kubelet node2} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.32&#34; in 284.243645ms&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:25 +0000 UTC - event for execpod8f26v: {kubelet node2} Created: Created container agnhost-container&#xA;Nov 22 02:29:42.397: INFO: At 2021-11-22 02:27:26 +0000 UTC - event for execpod8f26v: {kubelet node2} Started: Started container agnhost-container&#xA;Nov 22 02:29:42.400: INFO: POD                            NODE   PHASE    GRACE  CONDITIONS&#xA;Nov 22 02:29:42.400: INFO: execpod8f26v                   node2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:19 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:27 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:27 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:19 +0000 UTC  }]&#xA;Nov 22 02:29:42.400: INFO: nodeport-update-service-246qt  node2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:10 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:10 +0000 UTC  }]&#xA;Nov 22 02:29:42.400: INFO: nodeport-update-service-jgrp8  node2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:10 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:15 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:15 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-11-22 02:27:10 +0000 UTC  }]&#xA;Nov 22 02:29:42.400: INFO: &#xA;Nov 22 02:29:42.404: INFO: &#xA;Logging node info for node master1&#xA;Nov 22 02:29:42.407: INFO: Node Info: &amp;Node{ObjectMeta:{master1    2bbba583-0897-4d61-a459-4caacbb75db3 82310 0 2021-11-21 22:23:57 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/control-plane: node-role.kubernetes.io/master: node.kubernetes.io/exclude-from-external-load-balancers:] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.8.2 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubeadm Update v1 2021-11-21 22:24:00 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/control-plane&#34;:{},&#34;f:node-role.kubernetes.io/master&#34;:{},&#34;f:node.kubernetes.io/exclude-from-external-load-balancers&#34;:{}}}}} {flanneld Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {kubelet Update v1 2021-11-21 22:31:34 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}} {nfd-master Update v1 2021-11-21 22:34:13 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-11-21 22:29:42 +0000 UTC,LastTransitionTime:2021-11-21 22:29:42 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:40 +0000 UTC,LastTransitionTime:2021-11-21 22:23:55 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:40 +0000 UTC,LastTransitionTime:2021-11-21 22:23:55 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:40 +0000 UTC,LastTransitionTime:2021-11-21 22:23:55 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-11-22 02:29:40 +0000 UTC,LastTransitionTime:2021-11-21 22:26:53 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:ef2eedcc1bf54b2887fb39b8570e5ca6,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:7c111576-0e94-4e05-91e2-cb3834258298,KernelVersion:3.10.0-1160.45.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.11,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:724523485,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[sirot/netperf-latest@sha256:23929b922bb077cb341450fc1c90ae42e6f75da5d7ce61cd1880b08560a5ff85 sirot/netperf-latest:latest],SizeBytes:282025213,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[k8s.gcr.io/nfd/node-feature-discovery@sha256:74a1cbd82354f148277e20cdce25d57816e355a896bc67f67a0f722164b16945 k8s.gcr.io/nfd/node-feature-discovery:v0.8.2],SizeBytes:108486428,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[k8s.gcr.io/coredns/coredns@sha256:cc8fb77bc2a0541949d1d9320a641b82fd392b0d3d8145469ca4709ae769980e k8s.gcr.io/coredns/coredns:v1.8.0],SizeBytes:42454755,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[localhost:30500/tasextender@sha256:d8832dc123d295a3bf913b43c6f7207a7de223d86632f9215cfa92b5c97e6826 tasextender:latest localhost:30500/tasextender:0.4],SizeBytes:28910791,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[nginx@sha256:b92d3b942c8b84da889ac3dc6e83bd20ffb8cd2d8298eba92c8b0bf88d52f03e nginx:1.20.1-alpine],SizeBytes:22721538,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;Nov 22 02:29:42.407: INFO: &#xA;Logging kubelet events for node master1&#xA;Nov 22 02:29:42.409: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;Nov 22 02:29:42.430: INFO: node-exporter-v27cz started at 2021-11-21 22:39:21 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: kube-scheduler-master1 started at 2021-11-21 22:29:30 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: kube-proxy-4hrvq started at 2021-11-21 22:25:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;Nov 22 02:29:42.430: INFO: kube-multus-ds-amd64-bz55g started at 2021-11-21 22:26:58 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;Nov 22 02:29:42.430: INFO: container-registry-65d7c44b96-b5flj started at 2021-11-21 22:31:07 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: node-feature-discovery-controller-cff799f9f-55zvl started at 2021-11-21 22:34:07 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: kube-apiserver-master1 started at 2021-11-21 22:25:13 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: kube-controller-manager-master1 started at 2021-11-21 22:29:30 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;Nov 22 02:29:42.430: INFO: kube-flannel-djnzr started at 2021-11-21 22:26:48 +0000 UTC (1+1 container statuses recorded)&#xA;Nov 22 02:29:42.430: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;Nov 22 02:29:42.430: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;W1122 02:29:42.444860      28 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;Nov 22 02:29:42.524: INFO: &#xA;Latency metrics for node master1&#xA;Nov 22 02:29:42.524: INFO: &#xA;Logging node info for node master2&#xA;Nov 22 02:29:42.526: INFO: Node Info: &amp;Node{ObjectMeta:{master2    34181ea0-6ba6-41e9-b131-216ccaf91151 82293 0 2021-11-21 22:24:28 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/control-plane: node-role.kubernetes.io/master: node.kubernetes.io/exclude-from-external-load-balancers:] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubeadm Update v1 2021-11-21 22:24:29 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/control-plane&#34;:{},&#34;f:node-role.kubernetes.io/master&#34;:{},&#34;f:node.kubernetes.io/exclude-from-external-load-balancers&#34;:{}}}}} {flanneld Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {kubelet Update v1 2021-11-21 22:36:29 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-11-21 22:29:36 +0000 UTC,LastTransitionTime:2021-11-21 22:29:36 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:24:28 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:24:28 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:24:28 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:29:29 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:16378b49491c4207849d342feef6e088,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:18005186-998d-42f7-bb6b-754425f042ef,KernelVersion:3.10.0-1160.45.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.11,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:724523485,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[sirot/netperf-latest@sha256:23929b922bb077cb341450fc1c90ae42e6f75da5d7ce61cd1880b08560a5ff85 sirot/netperf-latest:latest],SizeBytes:282025213,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[quay.io/prometheus-operator/prometheus-operator@sha256:850c86bfeda4389bc9c757a9fd17ca5a090ea6b424968178d4467492cfa13921 quay.io/prometheus-operator/prometheus-operator:v0.44.1],SizeBytes:42617274,},ContainerImage{Names:[k8s.gcr.io/coredns/coredns@sha256:cc8fb77bc2a0541949d1d9320a641b82fd392b0d3d8145469ca4709ae769980e k8s.gcr.io/coredns/coredns:v1.8.0],SizeBytes:42454755,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;Nov 22 02:29:42.527: INFO: &#xA;Logging kubelet events for node master2&#xA;Nov 22 02:29:42.529: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;Nov 22 02:29:42.538: INFO: coredns-8474476ff8-8sq29 started at 2021-11-21 22:27:27 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;Nov 22 02:29:42.538: INFO: prometheus-operator-585ccfb458-hjlb7 started at 2021-11-21 22:39:13 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;Nov 22 02:29:42.538: INFO: kube-flannel-d4c7z started at 2021-11-21 22:26:48 +0000 UTC (1+1 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;Nov 22 02:29:42.538: INFO: kube-controller-manager-master2 started at 2021-11-21 22:29:29 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-controller-manager ready: true, restart count 3&#xA;Nov 22 02:29:42.538: INFO: kube-scheduler-master2 started at 2021-11-21 22:25:13 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-scheduler ready: true, restart count 3&#xA;Nov 22 02:29:42.538: INFO: kube-proxy-rvbqs started at 2021-11-21 22:25:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;Nov 22 02:29:42.538: INFO: kube-multus-ds-amd64-ldddk started at 2021-11-21 22:26:58 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;Nov 22 02:29:42.538: INFO: node-exporter-h9qzs started at 2021-11-21 22:39:21 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;Nov 22 02:29:42.538: INFO: kube-apiserver-master2 started at 2021-11-21 22:29:29 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.538: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;W1122 02:29:42.551810      28 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;Nov 22 02:29:42.616: INFO: &#xA;Latency metrics for node master2&#xA;Nov 22 02:29:42.616: INFO: &#xA;Logging node info for node master3&#xA;Nov 22 02:29:42.619: INFO: Node Info: &amp;Node{ObjectMeta:{master3    832a3406-d359-4476-8052-7ac28ab688e4 82319 0 2021-11-21 22:24:45 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/control-plane: node-role.kubernetes.io/master: node.kubernetes.io/exclude-from-external-load-balancers:] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubeadm Update v1 2021-11-21 22:24:46 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/control-plane&#34;:{},&#34;f:node-role.kubernetes.io/master&#34;:{},&#34;f:node.kubernetes.io/exclude-from-external-load-balancers&#34;:{}}}}} {flanneld Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {kubelet Update v1 2021-11-21 22:36:29 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-11-21 22:29:27 +0000 UTC,LastTransitionTime:2021-11-21 22:29:27 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:42 +0000 UTC,LastTransitionTime:2021-11-21 22:24:45 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:42 +0000 UTC,LastTransitionTime:2021-11-21 22:24:45 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:42 +0000 UTC,LastTransitionTime:2021-11-21 22:24:45 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-11-22 02:29:42 +0000 UTC,LastTransitionTime:2021-11-21 22:29:30 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31915f7485284e85a9f98d22e0e3e66d,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:821a678d-c5d5-4c70-a4e8-156f0093e267,KernelVersion:3.10.0-1160.45.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.11,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:724523485,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[k8s.gcr.io/coredns/coredns@sha256:cc8fb77bc2a0541949d1d9320a641b82fd392b0d3d8145469ca4709ae769980e k8s.gcr.io/coredns/coredns:v1.8.0],SizeBytes:42454755,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[aquasec/kube-bench@sha256:3544f6662feb73d36fdba35b17652e2fd73aae45bd4b60e76d7ab928220b3cc6 aquasec/kube-bench:0.3.1],SizeBytes:19301876,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;Nov 22 02:29:42.619: INFO: &#xA;Logging kubelet events for node master3&#xA;Nov 22 02:29:42.622: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;Nov 22 02:29:42.631: INFO: kube-apiserver-master3 started at 2021-11-21 22:24:46 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;Nov 22 02:29:42.631: INFO: kube-controller-manager-master3 started at 2021-11-21 22:32:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container kube-controller-manager ready: true, restart count 1&#xA;Nov 22 02:29:42.631: INFO: kube-scheduler-master3 started at 2021-11-21 22:29:30 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container kube-scheduler ready: true, restart count 1&#xA;Nov 22 02:29:42.631: INFO: kube-proxy-qj4jj started at 2021-11-21 22:25:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;Nov 22 02:29:42.631: INFO: kube-flannel-j5w8n started at 2021-11-21 22:26:48 +0000 UTC (1+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Init container install-cni ready: true, restart count 1&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;Nov 22 02:29:42.631: INFO: coredns-8474476ff8-n4cp8 started at 2021-11-21 22:27:23 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;Nov 22 02:29:42.631: INFO: dns-autoscaler-7df78bfcfb-4r8hz started at 2021-11-21 22:27:25 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;Nov 22 02:29:42.631: INFO: node-exporter-4m88h started at 2021-11-21 22:39:21 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;Nov 22 02:29:42.631: INFO: kube-multus-ds-amd64-h2qps started at 2021-11-21 22:26:58 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.631: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;W1122 02:29:42.646312      28 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;Nov 22 02:29:42.716: INFO: &#xA;Latency metrics for node master3&#xA;Nov 22 02:29:42.716: INFO: &#xA;Logging node info for node node1&#xA;Nov 22 02:29:42.719: INFO: Node Info: &amp;Node{ObjectMeta:{node1    9c5ba24f-75d8-4eb0-a15c-2525e54e2c6b 82294 0 2021-11-21 22:25:50 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.SSE4:true feature.node.kubernetes.io/cpu-cpuid.SSE42:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-cstate.enabled:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.status:active feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.45.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.SSE4,cpu-cpuid.SSE42,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-cstate.enabled,cpu-hardware_multithreading,cpu-pstate.status,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.8.2 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-11-21 22:25:50 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-11-21 22:25:50 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-11-21 22:34:13 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE4&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE42&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cstate.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.status&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-11-21 22:37:39 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-11-22 01:15:24 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269628928 0} {&lt;nil&gt;} 196552372Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884628480 0} {&lt;nil&gt;} 174692020Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-11-21 22:29:35 +0000 UTC,LastTransitionTime:2021-11-21 22:29:35 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:25:50 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:25:50 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:25:50 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-11-22 02:29:37 +0000 UTC,LastTransitionTime:2021-11-21 22:29:30 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:a59664461c3d486fa76a07928005cbdf,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:24c68613-09b6-4b8e-b7a0-2b6bedb352fa,KernelVersion:3.10.0-1160.45.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.11,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[opnfv/barometer-collectd@sha256:f30e965aa6195e6ac4ca2410f5a15e3704c92e4afa5208178ca22a7911975d66],SizeBytes:1075575763,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1003782921,},ContainerImage{Names:[localhost:30500/cmk@sha256:6fa0739c84dd43dd904c238b402991e6007a94b161969e6fda78120160df7f93 cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:724523485,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:db2475a1dbb2149508e5db31d7d77a75e6600d54be645f37681f03f2762169ba golang:alpine3.12],SizeBytes:301186719,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/jessie-dnsutils@sha256:702a992280fb7c3303e84a5801acbb4c9c7fcf48cffe0e9c8be3f0c60f74cf89 k8s.gcr.io/e2e-test-images/jessie-dnsutils:1.4],SizeBytes:253371792,},ContainerImage{Names:[grafana/grafana@sha256:ba39bf5131dcc0464134a3ff0e26e8c6380415249fa725e5f619176601255172 grafana/grafana:7.5.4],SizeBytes:203572842,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:b899dbd1b9017b9a379f76ce5b40eead01a62762c4f2057eacef945c3c22d210 quay.io/prometheus/prometheus:v2.22.1],SizeBytes:168344243,},ContainerImage{Names:[nginx@sha256:a05b0cdd4fc1be3b224ba9662ebdf98fe44c09c0c9215b45f84344c12867002e nginx:1.21.1],SizeBytes:133175493,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:716d2f68314c5c4ddd5ecdb45183fcb4ed8019015982c1321571f863989b70b0 k8s.gcr.io/e2e-test-images/httpd:2.4.39-1],SizeBytes:126894770,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50 k8s.gcr.io/e2e-test-images/httpd:2.4.38-1],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nautilus@sha256:1f36a24cfb5e0c3f725d7565a867c2384282fcbeccc77b07b423c9da95763a9a k8s.gcr.io/e2e-test-images/nautilus:1.4],SizeBytes:121748345,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[k8s.gcr.io/nfd/node-feature-discovery@sha256:74a1cbd82354f148277e20cdce25d57816e355a896bc67f67a0f722164b16945 k8s.gcr.io/nfd/node-feature-discovery:v0.8.2],SizeBytes:108486428,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter@sha256:2b09a571757a12c0245f2f1a74db4d1b9386ff901cf57f5ce48a0a682bd0e3af directxman12/k8s-prometheus-adapter:v0.8.2],SizeBytes:68230450,},ContainerImage{Names:[k8s.gcr.io/build-image/debian-iptables@sha256:d226f3fd5f293ff513f53573a40c069b89d57d42338a1045b493bf702ac6b1f6 k8s.gcr.io/build-image/debian-iptables:buster-v1.6.5],SizeBytes:60182158,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:fa923f38831d8c20c7c80135ff0b5a9c704569b87162df817911d36a4bc42d1f nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.2],SizeBytes:42686989,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nginx@sha256:503b7abb89e57383eba61cc8a9cb0b495ea575c516108f7d972a6ff6e1ab3c9b k8s.gcr.io/e2e-test-images/nginx:1.14-1],SizeBytes:16032814,},ContainerImage{Names:[quay.io/prometheus-operator/prometheus-config-reloader@sha256:4dee0fcf1820355ddd6986c1317b555693776c731315544a99d6cc59a7e34ce9 quay.io/prometheus-operator/prometheus-config-reloader:v0.44.1],SizeBytes:13433274,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:d9459083f962de6bd980ae6a05be2a4cf670df6a1d898157bceb420342bec280 alpine:3.12],SizeBytes:5592390,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/busybox@sha256:39e1e963e5310e9c313bad51523be012ede7b35bb9316517d19089a010356592 k8s.gcr.io/e2e-test-images/busybox:1.29-1],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;Nov 22 02:29:42.719: INFO: &#xA;Logging kubelet events for node node1&#xA;Nov 22 02:29:42.721: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;Nov 22 02:29:42.734: INFO: prometheus-k8s-0 started at 2021-11-21 22:39:32 +0000 UTC (0+4 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container config-reloader ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;Nov 22 02:29:42.734: INFO: up-down-2-frbjg started at 2021-11-22 02:27:39 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container up-down-2 ready: false, restart count 0&#xA;Nov 22 02:29:42.734: INFO: kube-flannel-cfzcv started at 2021-11-21 22:26:48 +0000 UTC (1+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;Nov 22 02:29:42.734: INFO: node-exporter-jj5rx started at 2021-11-21 22:39:21 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: kubernetes-metrics-scraper-5558854cb-kzhf7 started at 2021-11-21 22:27:27 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 2&#xA;Nov 22 02:29:42.734: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-9xds6 started at 2021-11-21 22:35:19 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: collectd-zmh78 started at 2021-11-21 22:43:10 +0000 UTC (0+3 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: nginx-proxy-node1 started at 2021-11-21 22:25:50 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;Nov 22 02:29:42.734: INFO: kube-multus-ds-amd64-wcr4n started at 2021-11-21 22:26:58 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;Nov 22 02:29:42.734: INFO: up-down-2-hjwtv started at 2021-11-22 02:27:39 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container up-down-2 ready: false, restart count 0&#xA;Nov 22 02:29:42.734: INFO: kube-proxy-mb5cq started at 2021-11-21 22:25:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;Nov 22 02:29:42.734: INFO: up-down-3-h8nd6 started at 2021-11-22 02:28:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container up-down-3 ready: false, restart count 0&#xA;Nov 22 02:29:42.734: INFO: cmk-7wvgm started at 2021-11-21 22:38:17 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: up-down-3-qhmbh started at 2021-11-22 02:28:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container up-down-3 ready: false, restart count 0&#xA;Nov 22 02:29:42.734: INFO: node-feature-discovery-worker-lkpb8 started at 2021-11-21 22:34:07 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;Nov 22 02:29:42.734: INFO: cmk-init-discover-node1-brwt6 started at 2021-11-21 22:37:36 +0000 UTC (0+3 container statuses recorded)&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container discover ready: false, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container init ready: false, restart count 0&#xA;Nov 22 02:29:42.734: INFO: &#x9;Container install ready: false, restart count 0&#xA;W1122 02:29:42.749064      28 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;Nov 22 02:29:43.086: INFO: &#xA;Latency metrics for node node1&#xA;Nov 22 02:29:43.086: INFO: &#xA;Logging node info for node node2&#xA;Nov 22 02:29:43.091: INFO: Node Info: &amp;Node{ObjectMeta:{node2    8239d97c-7f1c-4b50-a4b9-ee9c0135b86a 82288 0 2021-11-21 22:25:50 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.SSE4:true feature.node.kubernetes.io/cpu-cpuid.SSE42:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-cstate.enabled:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.status:active feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.45.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:null flannel.alpha.coreos.com/backend-type:host-gw flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.SSE4,cpu-cpuid.SSE42,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-cstate.enabled,cpu-hardware_multithreading,cpu-pstate.status,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.8.2 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-11-21 22:25:50 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-11-21 22:25:50 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-11-21 22:26:53 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-11-21 22:34:14 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE4&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.SSE42&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cstate.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.status&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-11-21 22:38:01 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-11-22 01:15:18 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:allocatable&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:example.com/fakecpu&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{}},&#34;f:capacity&#34;:{&#34;f:ephemeral-storage&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{}},&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{}}},&#34;f:images&#34;:{}}}} {e2e.test Update v1 2021-11-22 01:16:17 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:example.com/fakecpu&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-11-21 22:29:33 +0000 UTC,LastTransitionTime:2021-11-21 22:29:33 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:35 +0000 UTC,LastTransitionTime:2021-11-21 22:25:50 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:35 +0000 UTC,LastTransitionTime:2021-11-21 22:25:50 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-11-22 02:29:35 +0000 UTC,LastTransitionTime:2021-11-21 22:25:50 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-11-22 02:29:35 +0000 UTC,LastTransitionTime:2021-11-21 22:29:30 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:f7ac7a5a4fa14ecb963cd8859464e44b,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:4c94afbc-9699-4a1a-a85f-52972142831b,KernelVersion:3.10.0-1160.45.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://20.10.11,KubeletVersion:v1.21.1,KubeProxyVersion:v1.21.1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[opnfv/barometer-collectd@sha256:f30e965aa6195e6ac4ca2410f5a15e3704c92e4afa5208178ca22a7911975d66],SizeBytes:1075575763,},ContainerImage{Names:[localhost:30500/cmk@sha256:6fa0739c84dd43dd904c238b402991e6007a94b161969e6fda78120160df7f93 localhost:30500/cmk:v1.5.1],SizeBytes:724523485,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:724523485,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[aquasec/kube-hunter@sha256:2be6820bc1d7e0f57193a9a27d5a3e16b2fd93c53747b03ce8ca48c6fc323781 aquasec/kube-hunter:0.3.1],SizeBytes:347611549,},ContainerImage{Names:[sirot/netperf-latest@sha256:23929b922bb077cb341450fc1c90ae42e6f75da5d7ce61cd1880b08560a5ff85 sirot/netperf-latest:latest],SizeBytes:282025213,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:b9217b835cdcb33853f50a9cf13617ee0f8b887c508c5ac5110720de154914e4 kubernetesui/dashboard-amd64:v2.2.0],SizeBytes:225135791,},ContainerImage{Names:[nginx@sha256:a05b0cdd4fc1be3b224ba9662ebdf98fe44c09c0c9215b45f84344c12867002e nginx:1.21.1],SizeBytes:133175493,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:53af05c2a6cddd32cebf5856f71994f5d41ef2a62824b87f140f2087f91e4a38 k8s.gcr.io/kube-proxy:v1.21.1],SizeBytes:130788187,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:716d2f68314c5c4ddd5ecdb45183fcb4ed8019015982c1321571f863989b70b0 k8s.gcr.io/e2e-test-images/httpd:2.4.39-1],SizeBytes:126894770,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:758db666ac7028534dba72e7e9bb1e57bb81b8196f976f7a5cc351ef8b3529e1 k8s.gcr.io/e2e-test-images/agnhost:2.32],SizeBytes:125930239,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:53a13cd1588391888c5a8ac4cef13d3ee6d229cd904038936731af7131d193a9 k8s.gcr.io/kube-apiserver:v1.21.1],SizeBytes:125612423,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/httpd@sha256:b913fa234cc3473cfe16e937d106b455a7609f927f59031c81aca791e2689b50 k8s.gcr.io/e2e-test-images/httpd:2.4.38-1],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nautilus@sha256:1f36a24cfb5e0c3f725d7565a867c2384282fcbeccc77b07b423c9da95763a9a k8s.gcr.io/e2e-test-images/nautilus:1.4],SizeBytes:121748345,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:3daf9c9f9fe24c3a7b92ce864ef2d8d610c84124cc7d98e68fdbe94038337228 k8s.gcr.io/kube-controller-manager:v1.21.1],SizeBytes:119825302,},ContainerImage{Names:[k8s.gcr.io/nfd/node-feature-discovery@sha256:74a1cbd82354f148277e20cdce25d57816e355a896bc67f67a0f722164b16945 k8s.gcr.io/nfd/node-feature-discovery:v0.8.2],SizeBytes:108486428,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/sample-apiserver@sha256:e7fddbaac4c3451da2365ab90bad149d32f11409738034e41e0f460927f7c276 k8s.gcr.io/e2e-test-images/sample-apiserver:1.17.4],SizeBytes:58172101,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:a8c4084db3b381f0806ea563c7ec842cc3604c57722a916c91fb59b00ff67d63 k8s.gcr.io/kube-scheduler:v1.21.1],SizeBytes:50635642,},ContainerImage{Names:[quay.io/brancz/kube-rbac-proxy@sha256:05e15e1164fd7ac85f5702b3f87ef548f4e00de3a79e6c4a6a34c92035497a9a quay.io/brancz/kube-rbac-proxy:v0.8.0],SizeBytes:48952053,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/regression-issue-74839@sha256:b4f1d8d61bdad84bd50442d161d5460e4019d53e989b64220fdbc62fc87d76bf k8s.gcr.io/e2e-test-images/regression-issue-74839:1.2],SizeBytes:44576952,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:fa923f38831d8c20c7c80135ff0b5a9c704569b87162df817911d36a4bc42d1f localhost:30500/sriov-device-plugin:v3.3.2],SizeBytes:42686989,},ContainerImage{Names:[localhost:30500/tasextender@sha256:d8832dc123d295a3bf913b43c6f7207a7de223d86632f9215cfa92b5c97e6826 localhost:30500/tasextender:0.4],SizeBytes:28910791,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:cf66a6bbd573fd819ea09c72e21b528e9252d58d01ae13564a29749de1e48e0f quay.io/prometheus/node-exporter:v1.0.1],SizeBytes:26430341,},ContainerImage{Names:[aquasec/kube-bench@sha256:3544f6662feb73d36fdba35b17652e2fd73aae45bd4b60e76d7ab928220b3cc6 aquasec/kube-bench:0.3.1],SizeBytes:19301876,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nginx@sha256:503b7abb89e57383eba61cc8a9cb0b495ea575c516108f7d972a6ff6e1ab3c9b k8s.gcr.io/e2e-test-images/nginx:1.14-1],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/nonewprivs@sha256:8ac1264691820febacf3aea5d152cbde6d10685731ec14966a9401c6f47a68ac k8s.gcr.io/e2e-test-images/nonewprivs:1.3],SizeBytes:7107254,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/busybox@sha256:39e1e963e5310e9c313bad51523be012ede7b35bb9316517d19089a010356592 k8s.gcr.io/e2e-test-images/busybox:1.29-1],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:6c3835cab3980f11b83277305d0d736051c32b17606f5ec59f1dda67c9ba3810 k8s.gcr.io/pause:3.4.1],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;Nov 22 02:29:43.093: INFO: &#xA;Logging kubelet events for node node2&#xA;Nov 22 02:29:43.096: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;Nov 22 02:29:43.109: INFO: node-exporter-r2vkb started at 2021-11-21 22:39:21 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: kube-flannel-rdjt7 started at 2021-11-21 22:26:48 +0000 UTC (1+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;Nov 22 02:29:43.109: INFO: collectd-6t47m started at 2021-11-21 22:43:10 +0000 UTC (0+3 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: nodeport-update-service-246qt started at 2021-11-22 02:27:10 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container nodeport-update-service ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: cmk-prx26 started at 2021-11-21 22:38:18 +0000 UTC (0+2 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: up-down-3-smvvk started at 2021-11-22 02:28:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container up-down-3 ready: false, restart count 0&#xA;Nov 22 02:29:43.109: INFO: kubernetes-dashboard-785dcbb76d-wrkrj started at 2021-11-21 22:27:27 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 1&#xA;Nov 22 02:29:43.109: INFO: node-feature-discovery-worker-slrp4 started at 2021-11-21 22:34:07 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-9jdcq started at 2021-11-21 22:35:19 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: cmk-init-discover-node2-8jdqf started at 2021-11-21 22:37:56 +0000 UTC (0+3 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container discover ready: false, restart count 0&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container init ready: false, restart count 0&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container install ready: false, restart count 0&#xA;Nov 22 02:29:43.109: INFO: tas-telemetry-aware-scheduling-84ff454dfb-q64pf started at 2021-11-21 22:42:22 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: up-down-2-m6jtb started at 2021-11-22 02:27:38 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container up-down-2 ready: false, restart count 0&#xA;Nov 22 02:29:43.109: INFO: kube-proxy-5xb56 started at 2021-11-21 22:25:53 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;Nov 22 02:29:43.109: INFO: cmk-webhook-6c9d5f8578-8fxd8 started at 2021-11-21 22:38:18 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: nodeport-update-service-jgrp8 started at 2021-11-22 02:27:10 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container nodeport-update-service ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: nginx-proxy-node2 started at 2021-11-21 22:25:50 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;Nov 22 02:29:43.109: INFO: execpod8f26v started at 2021-11-22 02:27:19 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container agnhost-container ready: true, restart count 0&#xA;Nov 22 02:29:43.109: INFO: kube-multus-ds-amd64-6bg2m started at 2021-11-21 22:26:58 +0000 UTC (0+1 container statuses recorded)&#xA;Nov 22 02:29:43.109: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;W1122 02:29:43.121360      28 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;Nov 22 02:29:43.324: INFO: &#xA;Latency metrics for node node2&#xA;Nov 22 02:29:43.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;services-224&#34; for this suite.&#xA;[AfterEach] [sig-network] Services&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750&#xA;</system-out>
      </testcase>
  </testsuite>