<?xml version="1.0" encoding="UTF-8"?>
  <testsuite name="Kubernetes e2e suite" tests="3" failures="1" errors="0" time="156.147">
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] ESIPP [Slow] should work for type=LoadBalancer" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] HostPathType Socket [Slow] Should fail on mounting non-existent socket &#39;does-not-exist-socket&#39; when HostPathType is HostPathSocket" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)(allowExpansion)][sig-windows] volume-expand should resize volume when PVC is edited while pod is using it" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Volume Operations Storm [Feature:vsphere] should create pod with many volumes and verify no attach call fails" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: aws] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] NFSPersistentVolumes[Disruptive][Flaky] when kubelet restarts Should test that a file written to the mount before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Inline-volume (xfs)][Slow] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] EndpointSlice should create Endpoints and EndpointSlices for Pods matching a Service" classname="Kubernetes e2e suite" time="35.135553409"></testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (ntfs)][sig-windows] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: ceph][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (ntfs)][sig-windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] Downward API volume should provide container&#39;s memory request [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)][sig-windows] provisioning should provision storage with mount options" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI mock volume CSIStorageCapacity [Feature:CSIStorageCapacity] CSIStorageCapacity used, insufficient capacity" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: windows-gcepd] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should support existing directories when readOnly specified in the volumeSource" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: rbd][Feature:Volumes][Serial] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)][sig-windows] subPath should fail if subpath file is outside the volume [Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: gcepd] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Services should be rejected when no endpoints exist" classname="Kubernetes e2e suite" time="8.332891345"></testcase>
      <testcase name="[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow][LinuxOnly]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv written before kubelet restart is readable after restart." classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-storage] In-tree Volumes [Driver: iscsi][Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow]" classname="Kubernetes e2e suite" time="0">
          <skipped></skipped>
      </testcase>
      <testcase name="[sig-network] Networking Granular Checks: Services should function for pod-Service: udp" classname="Kubernetes e2e suite" time="112.397774557">
          <failure type="Failure">/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/networking.go:173&#xA;May 29 00:40:56.307: Failed to find expected responses:&#xA;Tries 34&#xA;Command curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;&#xA;retrieved map[]&#xA;expected map[netserver-0:{} netserver-1:{}]&#xA;&#xA;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/network/utils.go:285</failure>
          <system-out>[BeforeEach] [sig-network] Networking&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174&#xA;STEP: Creating a kubernetes client&#xA;May 29 00:39:04.232: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;STEP: Building a namespace api object, basename nettest&#xA;STEP: Waiting for a default service account to be provisioned in namespace&#xA;[BeforeEach] [sig-network] Networking&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/networking.go:94&#xA;STEP: Executing a successful http request from the external internet&#xA;[It] should function for pod-Service: udp&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/networking.go:173&#xA;STEP: Performing setup for networking test in namespace nettest-6329&#xA;STEP: creating a selector&#xA;STEP: Creating the service pods in kubernetes&#xA;May 29 00:39:04.359: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable&#xA;May 29 00:39:04.391: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 29 00:39:06.395: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true)&#xA;May 29 00:39:08.394: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 29 00:39:10.395: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 29 00:39:12.396: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 29 00:39:14.394: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 29 00:39:16.394: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 29 00:39:18.394: INFO: The status of Pod netserver-0 is Running (Ready = false)&#xA;May 29 00:39:20.395: INFO: The status of Pod netserver-0 is Running (Ready = true)&#xA;May 29 00:39:20.400: INFO: The status of Pod netserver-1 is Running (Ready = false)&#xA;May 29 00:39:22.404: INFO: The status of Pod netserver-1 is Running (Ready = false)&#xA;May 29 00:39:24.404: INFO: The status of Pod netserver-1 is Running (Ready = false)&#xA;May 29 00:39:26.403: INFO: The status of Pod netserver-1 is Running (Ready = false)&#xA;May 29 00:39:28.404: INFO: The status of Pod netserver-1 is Running (Ready = false)&#xA;May 29 00:39:30.403: INFO: The status of Pod netserver-1 is Running (Ready = true)&#xA;STEP: Creating test pods&#xA;STEP: Getting node addresses&#xA;May 29 00:39:36.425: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable&#xA;STEP: Creating the service on top of the pods in kubernetes&#xA;May 29 00:39:36.446: INFO: Service node-port-service in namespace nettest-6329 found.&#xA;May 29 00:39:36.459: INFO: Service session-affinity-service in namespace nettest-6329 found.&#xA;STEP: Waiting for NodePort service to expose endpoint&#xA;May 29 00:39:37.462: INFO: Waiting for amount of service:node-port-service endpoints to be 2&#xA;STEP: Waiting for Session Affinity service to expose endpoint&#xA;May 29 00:39:38.465: INFO: Waiting for amount of service:session-affinity-service endpoints to be 2&#xA;STEP: dialing(udp) test-container-pod --&gt; 10.233.25.107:90 (config.clusterIP)&#xA;May 29 00:39:38.471: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.233.25.107&amp;port=90&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:38.471: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:38.747: INFO: Waiting for responses: map[netserver-0:{}]&#xA;May 29 00:39:40.750: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.233.25.107&amp;port=90&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:40.750: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:41.076: INFO: Waiting for responses: map[netserver-0:{}]&#xA;May 29 00:39:43.080: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.233.25.107&amp;port=90&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:43.080: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:43.192: INFO: Waiting for responses: map[]&#xA;STEP: dialing(udp) test-container-pod --&gt; 10.10.190.207:30952 (nodeIP)&#xA;May 29 00:39:43.194: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:43.195: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:43.298: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:39:45.301: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:45.301: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:45.408: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:39:47.411: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:47.411: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:47.645: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:39:49.647: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:49.648: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:49.822: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:39:51.825: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:51.825: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:52.088: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:39:54.091: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:54.091: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:54.475: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:39:56.478: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:56.479: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:56.590: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:39:58.593: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:39:58.594: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:39:58.705: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:00.707: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:00.707: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:00.832: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:02.836: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:02.836: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:03.119: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:05.122: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:05.122: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:05.229: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:07.232: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:07.232: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:07.336: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:09.341: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:09.341: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:09.474: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:11.478: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:11.478: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:11.578: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:13.582: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:13.582: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:13.705: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:15.707: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:15.707: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:15.813: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:17.817: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:17.817: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:17.924: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:19.928: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:19.928: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:20.036: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:22.039: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:22.039: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:22.148: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:24.151: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:24.151: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:24.266: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:26.269: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:26.269: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:26.382: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:28.387: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:28.387: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:28.516: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:30.521: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:30.521: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:30.632: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:32.636: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:32.636: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:32.764: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:34.767: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:34.767: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:34.880: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:36.884: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:36.884: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:36.994: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:38.997: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:38.997: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:39.106: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:41.111: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:41.111: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:41.256: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:43.260: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:43.260: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:43.370: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:45.374: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:45.374: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:45.478: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:47.484: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:47.484: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:47.589: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:49.593: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:49.593: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:49.696: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:51.702: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:51.702: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:51.805: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:53.809: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;] Namespace:nettest-6329 PodName:test-container-pod ContainerName:webserver Stdin:&lt;nil&gt; CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}&#xA;May 29 00:40:53.809: INFO: &gt;&gt;&gt; kubeConfig: /root/.kube/config&#xA;May 29 00:40:53.994: INFO: Waiting for responses: map[netserver-0:{} netserver-1:{}]&#xA;May 29 00:40:55.994: INFO: &#xA;Output of kubectl describe pod nettest-6329/netserver-0:&#xA;&#xA;May 29 00:40:55.994: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=nettest-6329 describe pod netserver-0 --namespace=nettest-6329&#39;&#xA;May 29 00:40:56.143: INFO: stderr: &#34;&#34;&#xA;May 29 00:40:56.143: INFO: stdout: &#34;Name:         netserver-0\nNamespace:    nettest-6329\nPriority:     0\nNode:         node1/10.10.190.207\nStart Time:   Sat, 29 May 2021 00:39:04 +0000\nLabels:       selector-ba14ea58-daed-4ce3-a6db-e366bf0650e9=true\nAnnotations:  k8s.v1.cni.cncf.io/network-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.4.73\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;86:a0:d0:9a:b9:57\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              k8s.v1.cni.cncf.io/networks-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.4.73\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;86:a0:d0:9a:b9:57\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              kubernetes.io/psp: collectd\nStatus:       Running\nIP:           10.244.4.73\nIPs:\n  IP:  10.244.4.73\nContainers:\n  webserver:\n    Container ID:  docker://99855db97197bf84ad92081e67ff89badad786ef36b1f6f915ef55fad1bf3d9f\n    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.20\n    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0\n    Ports:         8080/TCP, 8081/UDP\n    Host Ports:    0/TCP, 0/UDP\n    Args:\n      netexec\n      --http-port=8080\n      --udp-port=8081\n    State:          Running\n      Started:      Sat, 29 May 2021 00:39:06 +0000\n    Ready:          True\n    Restart Count:  0\n    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Environment:    &lt;none&gt;\n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-bfqv2 (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-bfqv2:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-bfqv2\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  kubernetes.io/hostname=node1\nTolerations:     node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n                 node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n  Type    Reason          Age   From               Message\n  ----    ------          ----  ----               -------\n  Normal  Scheduled       112s  default-scheduler  Successfully assigned nettest-6329/netserver-0 to node1\n  Normal  AddedInterface  111s  multus             Add eth0 [10.244.4.73/24]\n  Normal  Pulling         111s  kubelet            Pulling image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.20\&#34;\n  Normal  Pulled          110s  kubelet            Successfully pulled image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.20\&#34; in 599.226481ms\n  Normal  Created         110s  kubelet            Created container webserver\n  Normal  Started         110s  kubelet            Started container webserver\n&#34;&#xA;May 29 00:40:56.143: INFO: Name:         netserver-0&#xA;Namespace:    nettest-6329&#xA;Priority:     0&#xA;Node:         node1/10.10.190.207&#xA;Start Time:   Sat, 29 May 2021 00:39:04 +0000&#xA;Labels:       selector-ba14ea58-daed-4ce3-a6db-e366bf0650e9=true&#xA;Annotations:  k8s.v1.cni.cncf.io/network-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.4.73&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;86:a0:d0:9a:b9:57&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              k8s.v1.cni.cncf.io/networks-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.4.73&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;86:a0:d0:9a:b9:57&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              kubernetes.io/psp: collectd&#xA;Status:       Running&#xA;IP:           10.244.4.73&#xA;IPs:&#xA;  IP:  10.244.4.73&#xA;Containers:&#xA;  webserver:&#xA;    Container ID:  docker://99855db97197bf84ad92081e67ff89badad786ef36b1f6f915ef55fad1bf3d9f&#xA;    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.20&#xA;    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0&#xA;    Ports:         8080/TCP, 8081/UDP&#xA;    Host Ports:    0/TCP, 0/UDP&#xA;    Args:&#xA;      netexec&#xA;      --http-port=8080&#xA;      --udp-port=8081&#xA;    State:          Running&#xA;      Started:      Sat, 29 May 2021 00:39:06 +0000&#xA;    Ready:          True&#xA;    Restart Count:  0&#xA;    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Environment:    &lt;none&gt;&#xA;    Mounts:&#xA;      /var/run/secrets/kubernetes.io/serviceaccount from default-token-bfqv2 (ro)&#xA;Conditions:&#xA;  Type              Status&#xA;  Initialized       True &#xA;  Ready             True &#xA;  ContainersReady   True &#xA;  PodScheduled      True &#xA;Volumes:&#xA;  default-token-bfqv2:&#xA;    Type:        Secret (a volume populated by a Secret)&#xA;    SecretName:  default-token-bfqv2&#xA;    Optional:    false&#xA;QoS Class:       BestEffort&#xA;Node-Selectors:  kubernetes.io/hostname=node1&#xA;Tolerations:     node.kubernetes.io/not-ready:NoExecute op=Exists for 300s&#xA;                 node.kubernetes.io/unreachable:NoExecute op=Exists for 300s&#xA;Events:&#xA;  Type    Reason          Age   From               Message&#xA;  ----    ------          ----  ----               -------&#xA;  Normal  Scheduled       112s  default-scheduler  Successfully assigned nettest-6329/netserver-0 to node1&#xA;  Normal  AddedInterface  111s  multus             Add eth0 [10.244.4.73/24]&#xA;  Normal  Pulling         111s  kubelet            Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34;&#xA;  Normal  Pulled          110s  kubelet            Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34; in 599.226481ms&#xA;  Normal  Created         110s  kubelet            Created container webserver&#xA;  Normal  Started         110s  kubelet            Started container webserver&#xA;&#xA;May 29 00:40:56.143: INFO: &#xA;Output of kubectl describe pod nettest-6329/netserver-1:&#xA;&#xA;May 29 00:40:56.143: INFO: Running &#39;/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=nettest-6329 describe pod netserver-1 --namespace=nettest-6329&#39;&#xA;May 29 00:40:56.306: INFO: stderr: &#34;&#34;&#xA;May 29 00:40:56.306: INFO: stdout: &#34;Name:         netserver-1\nNamespace:    nettest-6329\nPriority:     0\nNode:         node2/10.10.190.208\nStart Time:   Sat, 29 May 2021 00:39:04 +0000\nLabels:       selector-ba14ea58-daed-4ce3-a6db-e366bf0650e9=true\nAnnotations:  k8s.v1.cni.cncf.io/network-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.3.176\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;4a:61:15:61:46:9b\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              k8s.v1.cni.cncf.io/networks-status:\n                [{\n                    \&#34;name\&#34;: \&#34;default-cni-network\&#34;,\n                    \&#34;interface\&#34;: \&#34;eth0\&#34;,\n                    \&#34;ips\&#34;: [\n                        \&#34;10.244.3.176\&#34;\n                    ],\n                    \&#34;mac\&#34;: \&#34;4a:61:15:61:46:9b\&#34;,\n                    \&#34;default\&#34;: true,\n                    \&#34;dns\&#34;: {}\n                }]\n              kubernetes.io/psp: collectd\nStatus:       Running\nIP:           10.244.3.176\nIPs:\n  IP:  10.244.3.176\nContainers:\n  webserver:\n    Container ID:  docker://dc4a484eb6fb204c546ee76e68fa70bfc2efcc215a5dbd49251d675f1394e96d\n    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.20\n    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0\n    Ports:         8080/TCP, 8081/UDP\n    Host Ports:    0/TCP, 0/UDP\n    Args:\n      netexec\n      --http-port=8080\n      --udp-port=8081\n    State:          Running\n      Started:      Sat, 29 May 2021 00:39:10 +0000\n    Ready:          True\n    Restart Count:  0\n    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3\n    Environment:    &lt;none&gt;\n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-bfqv2 (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-bfqv2:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-bfqv2\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  kubernetes.io/hostname=node2\nTolerations:     node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n                 node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n  Type    Reason          Age   From               Message\n  ----    ------          ----  ----               -------\n  Normal  Scheduled       112s  default-scheduler  Successfully assigned nettest-6329/netserver-1 to node2\n  Normal  AddedInterface  108s  multus             Add eth0 [10.244.3.176/24]\n  Normal  Pulling         108s  kubelet            Pulling image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.20\&#34;\n  Normal  Pulled          108s  kubelet            Successfully pulled image \&#34;k8s.gcr.io/e2e-test-images/agnhost:2.20\&#34; in 944.439934ms\n  Normal  Created         107s  kubelet            Created container webserver\n  Normal  Started         106s  kubelet            Started container webserver\n&#34;&#xA;May 29 00:40:56.306: INFO: Name:         netserver-1&#xA;Namespace:    nettest-6329&#xA;Priority:     0&#xA;Node:         node2/10.10.190.208&#xA;Start Time:   Sat, 29 May 2021 00:39:04 +0000&#xA;Labels:       selector-ba14ea58-daed-4ce3-a6db-e366bf0650e9=true&#xA;Annotations:  k8s.v1.cni.cncf.io/network-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.3.176&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;4a:61:15:61:46:9b&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              k8s.v1.cni.cncf.io/networks-status:&#xA;                [{&#xA;                    &#34;name&#34;: &#34;default-cni-network&#34;,&#xA;                    &#34;interface&#34;: &#34;eth0&#34;,&#xA;                    &#34;ips&#34;: [&#xA;                        &#34;10.244.3.176&#34;&#xA;                    ],&#xA;                    &#34;mac&#34;: &#34;4a:61:15:61:46:9b&#34;,&#xA;                    &#34;default&#34;: true,&#xA;                    &#34;dns&#34;: {}&#xA;                }]&#xA;              kubernetes.io/psp: collectd&#xA;Status:       Running&#xA;IP:           10.244.3.176&#xA;IPs:&#xA;  IP:  10.244.3.176&#xA;Containers:&#xA;  webserver:&#xA;    Container ID:  docker://dc4a484eb6fb204c546ee76e68fa70bfc2efcc215a5dbd49251d675f1394e96d&#xA;    Image:         k8s.gcr.io/e2e-test-images/agnhost:2.20&#xA;    Image ID:      docker-pullable://k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0&#xA;    Ports:         8080/TCP, 8081/UDP&#xA;    Host Ports:    0/TCP, 0/UDP&#xA;    Args:&#xA;      netexec&#xA;      --http-port=8080&#xA;      --udp-port=8081&#xA;    State:          Running&#xA;      Started:      Sat, 29 May 2021 00:39:10 +0000&#xA;    Ready:          True&#xA;    Restart Count:  0&#xA;    Liveness:       http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Readiness:      http-get http://:8080/healthz delay=10s timeout=30s period=10s #success=1 #failure=3&#xA;    Environment:    &lt;none&gt;&#xA;    Mounts:&#xA;      /var/run/secrets/kubernetes.io/serviceaccount from default-token-bfqv2 (ro)&#xA;Conditions:&#xA;  Type              Status&#xA;  Initialized       True &#xA;  Ready             True &#xA;  ContainersReady   True &#xA;  PodScheduled      True &#xA;Volumes:&#xA;  default-token-bfqv2:&#xA;    Type:        Secret (a volume populated by a Secret)&#xA;    SecretName:  default-token-bfqv2&#xA;    Optional:    false&#xA;QoS Class:       BestEffort&#xA;Node-Selectors:  kubernetes.io/hostname=node2&#xA;Tolerations:     node.kubernetes.io/not-ready:NoExecute op=Exists for 300s&#xA;                 node.kubernetes.io/unreachable:NoExecute op=Exists for 300s&#xA;Events:&#xA;  Type    Reason          Age   From               Message&#xA;  ----    ------          ----  ----               -------&#xA;  Normal  Scheduled       112s  default-scheduler  Successfully assigned nettest-6329/netserver-1 to node2&#xA;  Normal  AddedInterface  108s  multus             Add eth0 [10.244.3.176/24]&#xA;  Normal  Pulling         108s  kubelet            Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34;&#xA;  Normal  Pulled          108s  kubelet            Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34; in 944.439934ms&#xA;  Normal  Created         107s  kubelet            Created container webserver&#xA;  Normal  Started         106s  kubelet            Started container webserver&#xA;&#xA;May 29 00:40:56.307: FAIL: Failed to find expected responses:&#xA;Tries 34&#xA;Command curl -g -q -s &#39;http://10.244.3.189:8080/dial?request=hostname&amp;protocol=udp&amp;host=10.10.190.207&amp;port=30952&amp;tries=1&#39;&#xA;retrieved map[]&#xA;expected map[netserver-0:{} netserver-1:{}]&#xA;&#xA;&#xA;Full Stack Trace&#xA;k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).DialFromContainer(0xc0019569c0, 0x4be8e5a, 0x3, 0x4bf21bf, 0x8, 0xc0011d4dc0, 0xc, 0xc0011d5bd0, 0xd, 0x1f90, ...)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/network/utils.go:285 +0x7db&#xA;k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).DialFromTestContainer(...)&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/network/utils.go:176&#xA;k8s.io/kubernetes/test/e2e/network.glob..func19.6.2()&#xA;&#x9;/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/networking.go:179 +0x3c5&#xA;k8s.io/kubernetes/test/e2e.RunE2ETests(0xc001b7f680)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e.go:130 +0x345&#xA;k8s.io/kubernetes/test/e2e.TestE2E(0xc001b7f680)&#xA;&#x9;_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/e2e_test.go:145 +0x2b&#xA;testing.tRunner(0xc001b7f680, 0x4de5140)&#xA;&#x9;/usr/local/go/src/testing/testing.go:1123 +0xef&#xA;created by testing.(*T).Run&#xA;&#x9;/usr/local/go/src/testing/testing.go:1168 +0x2b3&#xA;[AfterEach] [sig-network] Networking&#xA;  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175&#xA;STEP: Collecting events from namespace &#34;nettest-6329&#34;.&#xA;STEP: Found 18 events.&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:04 +0000 UTC - event for netserver-0: {default-scheduler } Scheduled: Successfully assigned nettest-6329/netserver-0 to node1&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:04 +0000 UTC - event for netserver-1: {default-scheduler } Scheduled: Successfully assigned nettest-6329/netserver-1 to node2&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:05 +0000 UTC - event for netserver-0: {multus } AddedInterface: Add eth0 [10.244.4.73/24]&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:05 +0000 UTC - event for netserver-0: {kubelet node1} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34;&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:06 +0000 UTC - event for netserver-0: {kubelet node1} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34; in 599.226481ms&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:06 +0000 UTC - event for netserver-0: {kubelet node1} Created: Created container webserver&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:06 +0000 UTC - event for netserver-0: {kubelet node1} Started: Started container webserver&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:08 +0000 UTC - event for netserver-1: {multus } AddedInterface: Add eth0 [10.244.3.176/24]&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:08 +0000 UTC - event for netserver-1: {kubelet node2} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34;&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:08 +0000 UTC - event for netserver-1: {kubelet node2} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34; in 944.439934ms&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:09 +0000 UTC - event for netserver-1: {kubelet node2} Created: Created container webserver&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:10 +0000 UTC - event for netserver-1: {kubelet node2} Started: Started container webserver&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:30 +0000 UTC - event for test-container-pod: {default-scheduler } Scheduled: Successfully assigned nettest-6329/test-container-pod to node2&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:32 +0000 UTC - event for test-container-pod: {multus } AddedInterface: Add eth0 [10.244.3.189/24]&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:32 +0000 UTC - event for test-container-pod: {kubelet node2} Pulling: Pulling image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34;&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:32 +0000 UTC - event for test-container-pod: {kubelet node2} Pulled: Successfully pulled image &#34;k8s.gcr.io/e2e-test-images/agnhost:2.20&#34; in 461.84842ms&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:33 +0000 UTC - event for test-container-pod: {kubelet node2} Created: Created container webserver&#xA;May 29 00:40:56.312: INFO: At 2021-05-29 00:39:33 +0000 UTC - event for test-container-pod: {kubelet node2} Started: Started container webserver&#xA;May 29 00:40:56.316: INFO: POD                 NODE   PHASE    GRACE  CONDITIONS&#xA;May 29 00:40:56.316: INFO: netserver-0         node1  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:04 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:19 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:19 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:04 +0000 UTC  }]&#xA;May 29 00:40:56.316: INFO: netserver-1         node2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:04 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:29 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:29 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:04 +0000 UTC  }]&#xA;May 29 00:40:56.316: INFO: test-container-pod  node2  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:30 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:33 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:33 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-05-29 00:39:30 +0000 UTC  }]&#xA;May 29 00:40:56.316: INFO: &#xA;May 29 00:40:56.320: INFO: &#xA;Logging node info for node master1&#xA;May 29 00:40:56.323: INFO: Node Info: &amp;Node{ObjectMeta:{master1   /api/v1/nodes/master1 0aa78934-442a-44a3-8c5c-f827e18dd3d7 97662 0 2021-05-28 19:56:25 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master1 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;0a:41:0b:9d:15:5a&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.202 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-28 19:56:25 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-28 19:56:26 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-28 19:59:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-28 19:59:09 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.0.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.0.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.0.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-28 20:02:03 +0000 UTC,LastTransitionTime:2021-05-28 20:02:03 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:54 +0000 UTC,LastTransitionTime:2021-05-28 19:56:25 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:54 +0000 UTC,LastTransitionTime:2021-05-28 19:56:25 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:54 +0000 UTC,LastTransitionTime:2021-05-28 19:56:25 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-29 00:40:54 +0000 UTC,LastTransitionTime:2021-05-28 19:59:09 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.202,},NodeAddress{Type:Hostname,Address:master1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:f7fb2c462cae4b9c990ab2e5c72f7816,SystemUUID:00ACFB60-0631-E711-906E-0017A4403562,BootID:24c06694-15ae-4da4-9143-144d98afdd8d,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726715672,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:9d07c391aeb1a9d02eb4343c113ed01825227c70c32b3cae861711f90191b0fd quay.io/coreos/kube-rbac-proxy:v0.4.1],SizeBytes:41317870,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[quay.io/coreos/prometheus-operator@sha256:a54e806fb27d2fb0251da4f3b2a3bb5320759af63a54a755788304775f2384a7 quay.io/coreos/prometheus-operator:v0.40.0],SizeBytes:38238457,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[registry@sha256:1cd9409a311350c3072fe510b52046f104416376c126a479cef9a4dfe692cf57 registry:2.7.0],SizeBytes:24191168,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:7f3d9945acdf5d86edd89b2b16fe1f6d63ba8bdb4cab50e66f9bce162df9e388 tas-controller:latest localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[nginx@sha256:a97eb9ecc708c8aa715ccfb5e9338f5456e4b65575daf304f108301f3b497314 nginx:1.19.2-alpine],SizeBytes:22052669,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:9af6075c93013910787a4e97973da6e0739a86dee1186d7965a5d00b1ac35636 tas-extender:latest localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:5577654,},ContainerImage{Names:[alpine@sha256:c0e9560cda118f9ec63ddefb4a173a2b2a0347082d7dff7dc14272e7841a5b5a alpine:3.12.1],SizeBytes:5573013,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 29 00:40:56.324: INFO: &#xA;Logging kubelet events for node master1&#xA;May 29 00:40:56.327: INFO: &#xA;Logging pods the kubelet thinks is on node master1&#xA;May 29 00:40:56.337: INFO: kube-proxy-994p2 started at 2021-05-28 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 29 00:40:56.337: INFO: kube-flannel-d54gm started at 2021-05-28 19:59:00 +0000 UTC (1+1 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 29 00:40:56.337: INFO: coredns-7677f9bb54-zb7h8 started at 2021-05-28 19:59:33 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May 29 00:40:56.337: INFO: node-exporter-9b7pq started at 2021-05-28 20:10:09 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: kube-scheduler-master1 started at 2021-05-28 19:57:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-scheduler ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: kube-apiserver-master1 started at 2021-05-28 20:05:21 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: kube-controller-manager-master1 started at 2021-05-28 19:57:39 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-controller-manager ready: true, restart count 2&#xA;May 29 00:40:56.337: INFO: kube-multus-ds-amd64-n9j8k started at 2021-05-28 19:59:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 29 00:40:56.337: INFO: docker-registry-docker-registry-56cbc7bc58-rbghz started at 2021-05-28 20:02:55 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container docker-registry ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: &#x9;Container nginx ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: prometheus-operator-5bb8cb9d8f-7wdtq started at 2021-05-28 20:10:02 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.337: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.337: INFO: &#x9;Container prometheus-operator ready: true, restart count 0&#xA;W0529 00:40:56.352202      26 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 29 00:40:56.382: INFO: &#xA;Latency metrics for node master1&#xA;May 29 00:40:56.382: INFO: &#xA;Logging node info for node master2&#xA;May 29 00:40:56.385: INFO: Node Info: &amp;Node{ObjectMeta:{master2   /api/v1/nodes/master2 b80f32b6-a396-4f09-a110-345a08d762ee 97610 0 2021-05-28 19:57:04 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master2 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;b2:be:c9:d8:cf:bb&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.203 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/master.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-28 19:57:04 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-28 19:57:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-28 19:59:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-28 19:59:09 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.1.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}} {nfd-master Update v1 2021-05-28 20:06:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/master.version&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.1.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.1.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234763776 0} {&lt;nil&gt;} 196518324Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324599808 0} {&lt;nil&gt;} 195629492Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-28 20:00:49 +0000 UTC,LastTransitionTime:2021-05-28 20:00:49 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 19:57:04 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 19:57:04 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 19:57:04 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 20:00:43 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.203,},NodeAddress{Type:Hostname,Address:master2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:2746caf91c53460599f165aa716150cd,SystemUUID:00A0DE53-E51D-E711-906E-0017A4403562,BootID:b63b522f-706f-4e28-a104-c73edcd04319,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726715672,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 29 00:40:56.385: INFO: &#xA;Logging kubelet events for node master2&#xA;May 29 00:40:56.388: INFO: &#xA;Logging pods the kubelet thinks is on node master2&#xA;May 29 00:40:56.396: INFO: kube-apiserver-master2 started at 2021-05-28 20:05:31 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 29 00:40:56.396: INFO: kube-controller-manager-master2 started at 2021-05-28 20:05:41 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container kube-controller-manager ready: true, restart count 3&#xA;May 29 00:40:56.396: INFO: kube-flannel-xvtkj started at 2021-05-28 19:59:00 +0000 UTC (1+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 29 00:40:56.396: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 29 00:40:56.396: INFO: kube-multus-ds-amd64-qjwcz started at 2021-05-28 19:59:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 29 00:40:56.396: INFO: node-feature-discovery-controller-5bf5c49849-n9ncl started at 2021-05-28 20:05:52 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container nfd-controller ready: true, restart count 0&#xA;May 29 00:40:56.396: INFO: kube-scheduler-master2 started at 2021-05-28 20:05:21 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container kube-scheduler ready: true, restart count 3&#xA;May 29 00:40:56.396: INFO: kube-proxy-jkbl8 started at 2021-05-28 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 29 00:40:56.396: INFO: dns-autoscaler-5b7b5c9b6f-r797x started at 2021-05-28 19:59:31 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container autoscaler ready: true, restart count 1&#xA;May 29 00:40:56.396: INFO: node-exporter-frch9 started at 2021-05-28 20:10:09 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.396: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.397: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;W0529 00:40:56.409396      26 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 29 00:40:56.432: INFO: &#xA;Latency metrics for node master2&#xA;May 29 00:40:56.432: INFO: &#xA;Logging node info for node master3&#xA;May 29 00:40:56.435: INFO: Node Info: &amp;Node{ObjectMeta:{master3   /api/v1/nodes/master3 301b0b5b-fc42-4c78-adb7-75baf6e0cc7e 97614 0 2021-05-28 19:57:14 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux kubernetes.io/arch:amd64 kubernetes.io/hostname:master3 kubernetes.io/os:linux node-role.kubernetes.io/master:] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;52:fa:ab:49:88:02&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.204 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kubelet Update v1 2021-05-28 19:57:14 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}} {kubeadm Update v1 2021-05-28 19:57:15 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}},&#34;f:labels&#34;:{&#34;f:node-role.kubernetes.io/master&#34;:{}}}}} {flanneld Update v1 2021-05-28 19:59:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {kube-controller-manager Update v1 2021-05-28 19:59:09 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.2.0/24\&#34;&#34;:{}},&#34;f:taints&#34;:{}}}}]},Spec:NodeSpec{PodCIDR:10.244.2.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:&lt;nil&gt;,},},ConfigSource:nil,PodCIDRs:[10.244.2.0/24],},Status:NodeStatus{Capacity:ResourceList{cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{201234767872 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{79550 -3} {&lt;nil&gt;} 79550m DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},memory: {{200324603904 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-28 20:02:12 +0000 UTC,LastTransitionTime:2021-05-28 20:02:12 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:51 +0000 UTC,LastTransitionTime:2021-05-28 19:57:14 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:51 +0000 UTC,LastTransitionTime:2021-05-28 19:57:14 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:51 +0000 UTC,LastTransitionTime:2021-05-28 19:57:14 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-29 00:40:51 +0000 UTC,LastTransitionTime:2021-05-28 19:59:09 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.204,},NodeAddress{Type:Hostname,Address:master3,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:a0cb6c0eb1d842469076fff344213c13,SystemUUID:008B1444-141E-E711-906E-0017A4403562,BootID:c6adcff4-8bf7-40d7-9d14-54b1c6a87bc8,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726715672,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[quay.io/coreos/etcd@sha256:04833b601fa130512450afa45c4fe484fee1293634f34c7ddc231bd193c74017 quay.io/coreos/etcd:v3.4.13],SizeBytes:83790470,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[k8s.gcr.io/coredns@sha256:73ca82b4ce829766d4f1f10947c3a338888f876fbed0540dc849c89ff256e90c k8s.gcr.io/coredns:1.7.0],SizeBytes:45227747,},ContainerImage{Names:[k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64@sha256:dce43068853ad396b0fb5ace9a56cc14114e31979e241342d12d04526be1dfcc k8s.gcr.io/cpa/cluster-proportional-autoscaler-amd64:1.8.3],SizeBytes:40647382,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 29 00:40:56.435: INFO: &#xA;Logging kubelet events for node master3&#xA;May 29 00:40:56.439: INFO: &#xA;Logging pods the kubelet thinks is on node master3&#xA;May 29 00:40:56.448: INFO: kube-controller-manager-master3 started at 2021-05-28 20:06:02 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Container kube-controller-manager ready: true, restart count 1&#xA;May 29 00:40:56.448: INFO: kube-scheduler-master3 started at 2021-05-28 20:01:23 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Container kube-scheduler ready: true, restart count 1&#xA;May 29 00:40:56.448: INFO: kube-proxy-t5bh6 started at 2021-05-28 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Container kube-proxy ready: true, restart count 1&#xA;May 29 00:40:56.448: INFO: kube-multus-ds-amd64-wqgf7 started at 2021-05-28 19:59:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 29 00:40:56.448: INFO: coredns-7677f9bb54-8v554 started at 2021-05-28 19:59:28 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Container coredns ready: true, restart count 1&#xA;May 29 00:40:56.448: INFO: node-exporter-w42s5 started at 2021-05-28 20:10:09 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.448: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 29 00:40:56.448: INFO: kube-apiserver-master3 started at 2021-05-28 20:05:21 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Container kube-apiserver ready: true, restart count 0&#xA;May 29 00:40:56.448: INFO: kube-flannel-zrskq started at 2021-05-28 19:59:00 +0000 UTC (1+1 container statuses recorded)&#xA;May 29 00:40:56.448: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 29 00:40:56.448: INFO: &#x9;Container kube-flannel ready: true, restart count 1&#xA;W0529 00:40:56.459656      26 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 29 00:40:56.489: INFO: &#xA;Latency metrics for node master3&#xA;May 29 00:40:56.489: INFO: &#xA;Logging node info for node node1&#xA;May 29 00:40:56.492: INFO: Node Info: &amp;Node{ObjectMeta:{node1   /api/v1/nodes/node1 43e51cb4-5acb-42b5-8f26-cd5e977f3829 97611 0 2021-05-28 19:58:22 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node1 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;d2:9d:b7:73:58:07&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.207 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-28 19:58:22 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.4.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-28 19:58:22 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-28 19:59:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-28 20:06:07 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-28 20:08:35 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {e2e.test Update v1 2021-05-28 22:27:50 +0000 UTC FieldsV1 {&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:example.com/fakecpu&#34;:{}}}}} {kubelet Update v1 2021-05-28 22:35:57 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:example.com/fakecpu&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.4.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.4.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269628928 0} {&lt;nil&gt;} 196552372Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},example.com/fakecpu: {{1 3} {&lt;nil&gt;} 1k DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884628480 0} {&lt;nil&gt;} 174692020Ki BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-28 20:01:58 +0000 UTC,LastTransitionTime:2021-05-28 20:01:58 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 19:58:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 19:58:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 19:58:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-29 00:40:49 +0000 UTC,LastTransitionTime:2021-05-28 19:59:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.207,},NodeAddress{Type:Hostname,Address:node1,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:abe6e95dbfa24a9abd34d8fa2abe7655,SystemUUID:00CDA902-D022-E711-906E-0017A4403562,BootID:17719d1f-7df5-4d95-81f3-7d3ac5110ba2,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:d731a0fc49b9ad6125b8d5dcb29da2b60bc940b48eacb6f5a9eb2a55c10598db localhost:30500/barometer-collectd:stable],SizeBytes:1464395058,},ContainerImage{Names:[&lt;none&gt;@&lt;none&gt; &lt;none&gt;:&lt;none&gt;],SizeBytes:1002495332,},ContainerImage{Names:[opnfv/barometer-collectd@sha256:ed5c574f653e2a39e784ff322033a2319aafde7366c803a88f20f7a2a8bc1efb opnfv/barometer-collectd:stable],SizeBytes:825413035,},ContainerImage{Names:[localhost:30500/cmk@sha256:97953d03767e4c2eb5d156394aeaf4bb0b74f3fd1ad08c303cb7561e272a00ff cmk:v1.5.1 localhost:30500/cmk:v1.5.1],SizeBytes:726715672,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[golang@sha256:aa24a0a337084e0747e7c8e97e1131270ae38150e691314f1fa19f4b2f9093c0 golang:alpine3.12],SizeBytes:301156062,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[k8s.gcr.io/etcd@sha256:4ad90a11b55313b182afc186b9876c8e891531b8db4c9bf1541953021618d0e2 k8s.gcr.io/etcd:3.4.13-0],SizeBytes:253392289,},ContainerImage{Names:[kubernetesui/dashboard-amd64@sha256:3af248961c56916aeca8eb4000c15d6cf6a69641ea92f0540865bb37b495932f kubernetesui/dashboard-amd64:v2.1.0],SizeBytes:225733746,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[grafana/grafana@sha256:89304bc2335f4976618548d7b93d165ed67369d3a051d2f627fc4e0aa3d0aff1 grafana/grafana:7.1.0],SizeBytes:179601493,},ContainerImage{Names:[quay.io/prometheus/prometheus@sha256:d4ba4dd1a9ebb90916d0bfed3c204adcb118ed24546bf8dd2e6b30fc0fd2009e quay.io/prometheus/prometheus:v2.20.0],SizeBytes:144886595,},ContainerImage{Names:[nginx@sha256:df13abe416e37eb3db4722840dd479b00ba193ac6606e7902331dcea50f4f1f2 nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/sample-apiserver@sha256:ff02aacd9766d597883fabafc7ad604c719a57611db1bcc1564c69a45b000a55 gcr.io/kubernetes-e2e-test-images/sample-apiserver:1.17],SizeBytes:60684726,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[directxman12/k8s-prometheus-adapter-amd64@sha256:b63dc612e3cb73f79d2401a4516f794f9f0a83002600ca72e675e41baecff437 directxman12/k8s-prometheus-adapter-amd64:v0.6.0],SizeBytes:53267842,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:2bec7a43da8efe70cb7cb14020a6b10aecd02c87e020d394de84e6807e2cf620 nfvpe/sriov-device-plugin:latest localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392623,},ContainerImage{Names:[kubernetesui/metrics-scraper@sha256:1f977343873ed0e2efd4916a6b2f3075f310ff6fe42ee098f54fc58aa7a28ab7 kubernetesui/metrics-scraper:v1.0.6],SizeBytes:34548789,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:7f3d9945acdf5d86edd89b2b16fe1f6d63ba8bdb4cab50e66f9bce162df9e388 localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:9af6075c93013910787a4e97973da6e0739a86dee1186d7965a5d00b1ac35636 localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[quay.io/coreos/prometheus-config-reloader@sha256:c679a143b24b7731ad1577a9865aa3805426cbf1b25e30807b951dff68466ffd quay.io/coreos/prometheus-config-reloader:v0.40.0],SizeBytes:10131705,},ContainerImage{Names:[jimmidyson/configmap-reload@sha256:d107c7a235c266273b1c3502a391fec374430e5625539403d0de797fa9c556a2 jimmidyson/configmap-reload:v0.3.0],SizeBytes:9700438,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[alpine@sha256:36553b10a4947067b9fbb7d532951066293a68eae893beba1d9235f7d11a20ad alpine:3.12],SizeBytes:5581415,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 29 00:40:56.493: INFO: &#xA;Logging kubelet events for node node1&#xA;May 29 00:40:56.497: INFO: &#xA;Logging pods the kubelet thinks is on node node1&#xA;May 29 00:40:56.511: INFO: cmk-jhzjr started at 2021-05-28 20:09:15 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: startup-script started at 2021-05-29 00:39:55 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container startup-script ready: false, restart count 0&#xA;May 29 00:40:56.511: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-zk2pt started at 2021-05-28 20:06:47 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: netserver-0 started at 2021-05-29 00:39:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: node-exporter-khdpg started at 2021-05-28 20:10:09 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: collectd-qw9nd started at 2021-05-28 20:16:29 +0000 UTC (0+3 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: &#x9;Container collectd-exporter ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: cmk-webhook-6c9d5f8578-kt8bp started at 2021-05-29 00:29:43 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container cmk-webhook ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: kube-flannel-2tjjt started at 2021-05-28 19:59:00 +0000 UTC (1+1 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Init container install-cni ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 29 00:40:56.511: INFO: node-feature-discovery-worker-5x4qg started at 2021-05-28 20:05:52 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 29 00:40:56.511: INFO: tas-telemetry-aware-scheduling-575ccbc9d4-6wq28 started at 2021-05-29 00:29:43 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.511: INFO: &#x9;Container tas-controller ready: true, restart count 0&#xA;May 29 00:40:56.512: INFO: &#x9;Container tas-extender ready: true, restart count 0&#xA;May 29 00:40:56.512: INFO: kubernetes-dashboard-86c6f9df5b-c5sbq started at 2021-05-28 19:59:33 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.512: INFO: &#x9;Container kubernetes-dashboard ready: true, restart count 2&#xA;May 29 00:40:56.512: INFO: prometheus-k8s-0 started at 2021-05-28 20:10:26 +0000 UTC (0+5 container statuses recorded)&#xA;May 29 00:40:56.512: INFO: &#x9;Container custom-metrics-apiserver ready: true, restart count 0&#xA;May 29 00:40:56.512: INFO: &#x9;Container grafana ready: true, restart count 0&#xA;May 29 00:40:56.512: INFO: &#x9;Container prometheus ready: true, restart count 1&#xA;May 29 00:40:56.512: INFO: &#x9;Container prometheus-config-reloader ready: true, restart count 0&#xA;May 29 00:40:56.512: INFO: &#x9;Container rules-configmap-reloader ready: true, restart count 0&#xA;May 29 00:40:56.512: INFO: cmk-init-discover-node1-rvqxm started at 2021-05-28 20:08:32 +0000 UTC (0+3 container statuses recorded)&#xA;May 29 00:40:56.512: INFO: &#x9;Container discover ready: false, restart count 0&#xA;May 29 00:40:56.512: INFO: &#x9;Container init ready: false, restart count 0&#xA;May 29 00:40:56.512: INFO: &#x9;Container install ready: false, restart count 0&#xA;May 29 00:40:56.512: INFO: kube-proxy-lsngv started at 2021-05-28 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.512: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 29 00:40:56.512: INFO: kube-multus-ds-amd64-x7826 started at 2021-05-28 19:59:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.512: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 29 00:40:56.512: INFO: kubernetes-metrics-scraper-678c97765c-wblkm started at 2021-05-28 19:59:33 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.512: INFO: &#x9;Container kubernetes-metrics-scraper ready: true, restart count 1&#xA;May 29 00:40:56.512: INFO: nginx-proxy-node1 started at 2021-05-28 20:05:21 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.512: INFO: &#x9;Container nginx-proxy ready: true, restart count 1&#xA;W0529 00:40:56.526469      26 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 29 00:40:56.558: INFO: &#xA;Latency metrics for node node1&#xA;May 29 00:40:56.558: INFO: &#xA;Logging node info for node node2&#xA;May 29 00:40:56.562: INFO: Node Info: &amp;Node{ObjectMeta:{node2   /api/v1/nodes/node2 3cc89580-b568-4c82-bd1f-200d0823da3b 97605 0 2021-05-28 19:58:22 +0000 UTC &lt;nil&gt; &lt;nil&gt; map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/os:linux cmk.intel.com/cmk-node:true feature.node.kubernetes.io/cpu-cpuid.ADX:true feature.node.kubernetes.io/cpu-cpuid.AESNI:true feature.node.kubernetes.io/cpu-cpuid.AVX:true feature.node.kubernetes.io/cpu-cpuid.AVX2:true feature.node.kubernetes.io/cpu-cpuid.AVX512BW:true feature.node.kubernetes.io/cpu-cpuid.AVX512CD:true feature.node.kubernetes.io/cpu-cpuid.AVX512DQ:true feature.node.kubernetes.io/cpu-cpuid.AVX512F:true feature.node.kubernetes.io/cpu-cpuid.AVX512VL:true feature.node.kubernetes.io/cpu-cpuid.FMA3:true feature.node.kubernetes.io/cpu-cpuid.HLE:true feature.node.kubernetes.io/cpu-cpuid.IBPB:true feature.node.kubernetes.io/cpu-cpuid.MPX:true feature.node.kubernetes.io/cpu-cpuid.RTM:true feature.node.kubernetes.io/cpu-cpuid.STIBP:true feature.node.kubernetes.io/cpu-cpuid.VMX:true feature.node.kubernetes.io/cpu-hardware_multithreading:true feature.node.kubernetes.io/cpu-pstate.turbo:true feature.node.kubernetes.io/cpu-rdt.RDTCMT:true feature.node.kubernetes.io/cpu-rdt.RDTL3CA:true feature.node.kubernetes.io/cpu-rdt.RDTMBA:true feature.node.kubernetes.io/cpu-rdt.RDTMBM:true feature.node.kubernetes.io/cpu-rdt.RDTMON:true feature.node.kubernetes.io/kernel-config.NO_HZ:true feature.node.kubernetes.io/kernel-config.NO_HZ_FULL:true feature.node.kubernetes.io/kernel-selinux.enabled:true feature.node.kubernetes.io/kernel-version.full:3.10.0-1160.25.1.el7.x86_64 feature.node.kubernetes.io/kernel-version.major:3 feature.node.kubernetes.io/kernel-version.minor:10 feature.node.kubernetes.io/kernel-version.revision:0 feature.node.kubernetes.io/memory-numa:true feature.node.kubernetes.io/network-sriov.capable:true feature.node.kubernetes.io/network-sriov.configured:true feature.node.kubernetes.io/pci-0300_1a03.present:true feature.node.kubernetes.io/storage-nonrotationaldisk:true feature.node.kubernetes.io/system-os_release.ID:centos feature.node.kubernetes.io/system-os_release.VERSION_ID:7 feature.node.kubernetes.io/system-os_release.VERSION_ID.major:7 kubernetes.io/arch:amd64 kubernetes.io/hostname:node2 kubernetes.io/os:linux] map[flannel.alpha.coreos.com/backend-data:{&#34;VtepMAC&#34;:&#34;62:22:2c:ae:14:ae&#34;} flannel.alpha.coreos.com/backend-type:vxlan flannel.alpha.coreos.com/kube-subnet-manager:true flannel.alpha.coreos.com/public-ip:10.10.190.208 kubeadm.alpha.kubernetes.io/cri-socket:/var/run/dockershim.sock nfd.node.kubernetes.io/extended-resources: nfd.node.kubernetes.io/feature-labels:cpu-cpuid.ADX,cpu-cpuid.AESNI,cpu-cpuid.AVX,cpu-cpuid.AVX2,cpu-cpuid.AVX512BW,cpu-cpuid.AVX512CD,cpu-cpuid.AVX512DQ,cpu-cpuid.AVX512F,cpu-cpuid.AVX512VL,cpu-cpuid.FMA3,cpu-cpuid.HLE,cpu-cpuid.IBPB,cpu-cpuid.MPX,cpu-cpuid.RTM,cpu-cpuid.STIBP,cpu-cpuid.VMX,cpu-hardware_multithreading,cpu-pstate.turbo,cpu-rdt.RDTCMT,cpu-rdt.RDTL3CA,cpu-rdt.RDTMBA,cpu-rdt.RDTMBM,cpu-rdt.RDTMON,kernel-config.NO_HZ,kernel-config.NO_HZ_FULL,kernel-selinux.enabled,kernel-version.full,kernel-version.major,kernel-version.minor,kernel-version.revision,memory-numa,network-sriov.capable,network-sriov.configured,pci-0300_1a03.present,storage-nonrotationaldisk,system-os_release.ID,system-os_release.VERSION_ID,system-os_release.VERSION_ID.major nfd.node.kubernetes.io/worker.version:v0.7.0 node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] []  [{kube-controller-manager Update v1 2021-05-28 19:58:22 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:node.alpha.kubernetes.io/ttl&#34;:{}}},&#34;f:spec&#34;:{&#34;f:podCIDR&#34;:{},&#34;f:podCIDRs&#34;:{&#34;.&#34;:{},&#34;v:\&#34;10.244.3.0/24\&#34;&#34;:{}}}}} {kubeadm Update v1 2021-05-28 19:58:22 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:kubeadm.alpha.kubernetes.io/cri-socket&#34;:{}}}}} {flanneld Update v1 2021-05-28 19:59:05 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:flannel.alpha.coreos.com/backend-data&#34;:{},&#34;f:flannel.alpha.coreos.com/backend-type&#34;:{},&#34;f:flannel.alpha.coreos.com/kube-subnet-manager&#34;:{},&#34;f:flannel.alpha.coreos.com/public-ip&#34;:{}}},&#34;f:status&#34;:{&#34;f:conditions&#34;:{&#34;k:{\&#34;type\&#34;:\&#34;NetworkUnavailable\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}}}}} {nfd-master Update v1 2021-05-28 20:06:06 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;f:nfd.node.kubernetes.io/extended-resources&#34;:{},&#34;f:nfd.node.kubernetes.io/feature-labels&#34;:{},&#34;f:nfd.node.kubernetes.io/worker.version&#34;:{}},&#34;f:labels&#34;:{&#34;f:feature.node.kubernetes.io/cpu-cpuid.ADX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AESNI&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX2&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512BW&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512CD&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512DQ&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512F&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.AVX512VL&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.FMA3&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.HLE&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.IBPB&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.MPX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.RTM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.STIBP&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-cpuid.VMX&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-hardware_multithreading&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-pstate.turbo&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTCMT&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTL3CA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBA&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMBM&#34;:{},&#34;f:feature.node.kubernetes.io/cpu-rdt.RDTMON&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-config.NO_HZ_FULL&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-selinux.enabled&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.full&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.major&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.minor&#34;:{},&#34;f:feature.node.kubernetes.io/kernel-version.revision&#34;:{},&#34;f:feature.node.kubernetes.io/memory-numa&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.capable&#34;:{},&#34;f:feature.node.kubernetes.io/network-sriov.configured&#34;:{},&#34;f:feature.node.kubernetes.io/pci-0300_1a03.present&#34;:{},&#34;f:feature.node.kubernetes.io/storage-nonrotationaldisk&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID&#34;:{},&#34;f:feature.node.kubernetes.io/system-os_release.VERSION_ID.major&#34;:{}}}}} {Swagger-Codegen Update v1 2021-05-28 20:08:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:labels&#34;:{&#34;f:cmk.intel.com/cmk-node&#34;:{}}},&#34;f:status&#34;:{&#34;f:capacity&#34;:{&#34;f:cmk.intel.com/exclusive-cores&#34;:{}}}}} {kubelet Update v1 2021-05-28 22:35:58 +0000 UTC FieldsV1 {&#34;f:metadata&#34;:{&#34;f:annotations&#34;:{&#34;.&#34;:{},&#34;f:volumes.kubernetes.io/controller-managed-attach-detach&#34;:{}},&#34;f:labels&#34;:{&#34;.&#34;:{},&#34;f:beta.kubernetes.io/arch&#34;:{},&#34;f:beta.kubernetes.io/os&#34;:{},&#34;f:kubernetes.io/arch&#34;:{},&#34;f:kubernetes.io/hostname&#34;:{},&#34;f:kubernetes.io/os&#34;:{}}},&#34;f:status&#34;:{&#34;f:addresses&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;Hostname\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;InternalIP\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:address&#34;:{},&#34;f:type&#34;:{}}},&#34;f:allocatable&#34;:{&#34;.&#34;:{},&#34;f:cmk.intel.com/exclusive-cores&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:capacity&#34;:{&#34;.&#34;:{},&#34;f:cpu&#34;:{},&#34;f:ephemeral-storage&#34;:{},&#34;f:hugepages-1Gi&#34;:{},&#34;f:hugepages-2Mi&#34;:{},&#34;f:intel.com/intel_sriov_netdevice&#34;:{},&#34;f:memory&#34;:{},&#34;f:pods&#34;:{}},&#34;f:conditions&#34;:{&#34;.&#34;:{},&#34;k:{\&#34;type\&#34;:\&#34;DiskPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;MemoryPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;PIDPressure\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}},&#34;k:{\&#34;type\&#34;:\&#34;Ready\&#34;}&#34;:{&#34;.&#34;:{},&#34;f:lastHeartbeatTime&#34;:{},&#34;f:lastTransitionTime&#34;:{},&#34;f:message&#34;:{},&#34;f:reason&#34;:{},&#34;f:status&#34;:{},&#34;f:type&#34;:{}}},&#34;f:daemonEndpoints&#34;:{&#34;f:kubeletEndpoint&#34;:{&#34;f:Port&#34;:{}}},&#34;f:images&#34;:{},&#34;f:nodeInfo&#34;:{&#34;f:architecture&#34;:{},&#34;f:bootID&#34;:{},&#34;f:containerRuntimeVersion&#34;:{},&#34;f:kernelVersion&#34;:{},&#34;f:kubeProxyVersion&#34;:{},&#34;f:kubeletVersion&#34;:{},&#34;f:machineID&#34;:{},&#34;f:operatingSystem&#34;:{},&#34;f:osImage&#34;:{},&#34;f:systemUUID&#34;:{}}}}}]},Spec:NodeSpec{PodCIDR:10.244.3.0/24,DoNotUseExternalID:,ProviderID:,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.244.3.0/24],},Status:NodeStatus{Capacity:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{80 0} {&lt;nil&gt;} 80 DecimalSI},ephemeral-storage: {{450471260160 0} {&lt;nil&gt;} 439913340Ki BinarySI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{201269633024 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Allocatable:ResourceList{cmk.intel.com/exclusive-cores: {{3 0} {&lt;nil&gt;} 3 DecimalSI},cpu: {{77 0} {&lt;nil&gt;} 77 DecimalSI},ephemeral-storage: {{405424133473 0} {&lt;nil&gt;} 405424133473 DecimalSI},hugepages-1Gi: {{0 0} {&lt;nil&gt;} 0 DecimalSI},hugepages-2Mi: {{21474836480 0} {&lt;nil&gt;} 20Gi BinarySI},intel.com/intel_sriov_netdevice: {{4 0} {&lt;nil&gt;} 4 DecimalSI},memory: {{178884632576 0} {&lt;nil&gt;}  BinarySI},pods: {{110 0} {&lt;nil&gt;} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2021-05-28 20:01:05 +0000 UTC,LastTransitionTime:2021-05-28 20:01:05 +0000 UTC,Reason:FlannelIsUp,Message:Flannel is running on this node,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:47 +0000 UTC,LastTransitionTime:2021-05-28 19:58:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:47 +0000 UTC,LastTransitionTime:2021-05-28 19:58:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2021-05-29 00:40:47 +0000 UTC,LastTransitionTime:2021-05-28 19:58:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2021-05-29 00:40:47 +0000 UTC,LastTransitionTime:2021-05-28 19:59:04 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.10.190.208,},NodeAddress{Type:Hostname,Address:node2,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b2730c4b09814ab9a78e7bc62c820fbb,SystemUUID:80B3CD56-852F-E711-906E-0017A4403562,BootID:f1459072-d21d-46de-a5d9-46ec9349aae0,KernelVersion:3.10.0-1160.25.1.el7.x86_64,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://19.3.14,KubeletVersion:v1.19.8,KubeProxyVersion:v1.19.8,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[localhost:30500/barometer-collectd@sha256:d731a0fc49b9ad6125b8d5dcb29da2b60bc940b48eacb6f5a9eb2a55c10598db localhost:30500/barometer-collectd:stable],SizeBytes:1464395058,},ContainerImage{Names:[localhost:30500/cmk@sha256:97953d03767e4c2eb5d156394aeaf4bb0b74f3fd1ad08c303cb7561e272a00ff localhost:30500/cmk:v1.5.1],SizeBytes:726715672,},ContainerImage{Names:[cmk:v1.5.1],SizeBytes:726715672,},ContainerImage{Names:[centos/python-36-centos7@sha256:ac50754646f0d37616515fb30467d8743fb12954260ec36c9ecb5a94499447e0 centos/python-36-centos7:latest],SizeBytes:650061677,},ContainerImage{Names:[nfvpe/multus@sha256:ac1266b87ba44c09dc2a336f0d5dad968fccd389ce1944a85e87b32cd21f7224 nfvpe/multus:v3.4.2],SizeBytes:276587882,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/jessie-dnsutils@sha256:ad583e33cb284f7ef046673809b146ec4053cda19b54a85d2b180a86169715eb gcr.io/kubernetes-e2e-test-images/jessie-dnsutils:1.0],SizeBytes:195659796,},ContainerImage{Names:[nginx@sha256:df13abe416e37eb3db4722840dd479b00ba193ac6606e7902331dcea50f4f1f2 nginx:1.19],SizeBytes:133122553,},ContainerImage{Names:[httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060 httpd:2.4.38-alpine],SizeBytes:123781643,},ContainerImage{Names:[k8s.gcr.io/kube-apiserver@sha256:82e0ce4e1d08f3749d05c584fd60986197bfcdf9ce71d4666c71674221d53135 k8s.gcr.io/kube-apiserver:v1.19.8],SizeBytes:118813022,},ContainerImage{Names:[k8s.gcr.io/kube-proxy@sha256:8ed30419d9cf8965854f9ed501159e15deb30c42c3d2a60a278ae169320d140e k8s.gcr.io/kube-proxy:v1.19.8],SizeBytes:117674285,},ContainerImage{Names:[k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0 k8s.gcr.io/e2e-test-images/agnhost:2.20],SizeBytes:113869866,},ContainerImage{Names:[k8s.gcr.io/kube-controller-manager@sha256:2769005fb667dbb936009894d01fe35f5ce1bce45eee80a9ce3c139b9be4080e k8s.gcr.io/kube-controller-manager:v1.19.8],SizeBytes:110805342,},ContainerImage{Names:[gcr.io/k8s-staging-nfd/node-feature-discovery@sha256:5d116c2c340be665a2c8adc9aca7f91396bd5cbde4add4fdc8dab95d8db43425 gcr.io/k8s-staging-nfd/node-feature-discovery:v0.7.0],SizeBytes:108309584,},ContainerImage{Names:[quay.io/coreos/flannel@sha256:34860ea294a018d392e61936f19a7862d5e92039d196cac9176da14b2bbd0fe3 quay.io/coreos/flannel@sha256:ac5322604bcab484955e6dbc507f45a906bde79046667322e3918a8578ab08c8 quay.io/coreos/flannel:v0.13.0 quay.io/coreos/flannel:v0.13.0-amd64],SizeBytes:57156911,},ContainerImage{Names:[quay.io/coreos/kube-rbac-proxy@sha256:e10d1d982dd653db74ca87a1d1ad017bc5ef1aeb651bdea089debf16485b080b quay.io/coreos/kube-rbac-proxy:v0.5.0],SizeBytes:46626428,},ContainerImage{Names:[k8s.gcr.io/kube-scheduler@sha256:bb66135ce9a25ac405e43bbae6a2ac766e0efcac0a6a73ef9d1fbb4cf4732c9b k8s.gcr.io/kube-scheduler:v1.19.8],SizeBytes:46510430,},ContainerImage{Names:[localhost:30500/sriov-device-plugin@sha256:2bec7a43da8efe70cb7cb14020a6b10aecd02c87e020d394de84e6807e2cf620 localhost:30500/sriov-device-plugin:v3.3.1],SizeBytes:44392623,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonroot@sha256:4bd7ae247de5c988700233c5a4b55e804ffe90f8c66ae64853f1dae37b847213 gcr.io/kubernetes-e2e-test-images/nonroot:1.0],SizeBytes:42321438,},ContainerImage{Names:[quay.io/prometheus/node-exporter@sha256:a2f29256e53cc3e0b64d7a472512600b2e9410347d53cdc85b49f659c17e02ee quay.io/prometheus/node-exporter:v0.18.1],SizeBytes:22933477,},ContainerImage{Names:[localhost:30500/tas-controller@sha256:7f3d9945acdf5d86edd89b2b16fe1f6d63ba8bdb4cab50e66f9bce162df9e388 localhost:30500/tas-controller:0.1],SizeBytes:22922439,},ContainerImage{Names:[localhost:30500/tas-extender@sha256:9af6075c93013910787a4e97973da6e0739a86dee1186d7965a5d00b1ac35636 localhost:30500/tas-extender:0.1],SizeBytes:21320903,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/regression-issue-74839-amd64@sha256:3b36bd80b97c532a774e7f6246797b8575d97037982f353476c703ba6686c75c gcr.io/kubernetes-e2e-test-images/regression-issue-74839-amd64:1.0],SizeBytes:19227369,},ContainerImage{Names:[prom/collectd-exporter@sha256:73fbda4d24421bff3b741c27efc36f1b6fbe7c57c378d56d4ff78101cd556654],SizeBytes:17463681,},ContainerImage{Names:[nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 nginx:1.14-alpine],SizeBytes:16032814,},ContainerImage{Names:[gcr.io/google-samples/hello-go-gke@sha256:4ea9cd3d35f81fc91bdebca3fae50c180a1048be0613ad0f811595365040396e gcr.io/google-samples/hello-go-gke:1.0],SizeBytes:11443478,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nonewprivs@sha256:10066e9039219449fe3c81f38fe01928f87914150768ab81b62a468e51fa7411 gcr.io/kubernetes-e2e-test-images/nonewprivs:1.0],SizeBytes:6757579,},ContainerImage{Names:[appropriate/curl@sha256:027a0ad3c69d085fea765afca9984787b780c172cead6502fec989198b98d8bb appropriate/curl:edge],SizeBytes:5654234,},ContainerImage{Names:[gcr.io/kubernetes-e2e-test-images/nautilus@sha256:33a732d4c42a266912a5091598a0f07653c9134db4b8d571690d8afd509e0bfc gcr.io/kubernetes-e2e-test-images/nautilus:1.0],SizeBytes:4753501,},ContainerImage{Names:[gcr.io/authenticated-image-pulling/alpine@sha256:7ff177862cb50c602bfe81f805969412e619c054a2bbead977d0c276988aa4a0 gcr.io/authenticated-image-pulling/alpine:3.7],SizeBytes:4206620,},ContainerImage{Names:[busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796 busybox:1.29],SizeBytes:1154361,},ContainerImage{Names:[busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47 busybox:1.28],SizeBytes:1146369,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:927d98197ec1141a368550822d18fa1c60bdae27b78b0c004f705f548c07814f k8s.gcr.io/pause:3.2],SizeBytes:682696,},ContainerImage{Names:[k8s.gcr.io/pause@sha256:a319ac2280eb7e3a59e252e54b76327cb4a33cf8389053b0d78277f22bbca2fa k8s.gcr.io/pause:3.3],SizeBytes:682696,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},}&#xA;May 29 00:40:56.563: INFO: &#xA;Logging kubelet events for node node2&#xA;May 29 00:40:56.567: INFO: &#xA;Logging pods the kubelet thinks is on node node2&#xA;May 29 00:40:56.577: INFO: nginx-proxy-node2 started at 2021-05-28 20:05:21 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container nginx-proxy ready: true, restart count 2&#xA;May 29 00:40:56.577: INFO: kube-proxy-z5czn started at 2021-05-28 19:58:24 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container kube-proxy ready: true, restart count 2&#xA;May 29 00:40:56.577: INFO: sriov-net-dp-kube-sriov-device-plugin-amd64-mkc6p started at 2021-05-29 00:29:50 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container kube-sriovdp ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: node-feature-discovery-worker-2qfpd started at 2021-05-29 00:29:50 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container nfd-worker ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: boom-server started at 2021-05-29 00:39:49 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container boom-server ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: kube-flannel-d9wsg started at 2021-05-28 19:59:00 +0000 UTC (1+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Init container install-cni ready: true, restart count 2&#xA;May 29 00:40:56.577: INFO: &#x9;Container kube-flannel ready: true, restart count 2&#xA;May 29 00:40:56.577: INFO: kube-multus-ds-amd64-c9cj2 started at 2021-05-28 19:59:08 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container kube-multus ready: true, restart count 1&#xA;May 29 00:40:56.577: INFO: test-container-pod started at 2021-05-29 00:39:30 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: cmk-lbg6n started at 2021-05-29 00:29:50 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container nodereport ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: &#x9;Container reconcile ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: node-exporter-nsrbd started at 2021-05-29 00:29:50 +0000 UTC (0+2 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container kube-rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: &#x9;Container node-exporter ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: collectd-k6rzg started at 2021-05-29 00:30:20 +0000 UTC (0+3 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container collectd ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: &#x9;Container collectd-exporter ready: false, restart count 0&#xA;May 29 00:40:56.577: INFO: &#x9;Container rbac-proxy ready: true, restart count 0&#xA;May 29 00:40:56.577: INFO: netserver-1 started at 2021-05-29 00:39:04 +0000 UTC (0+1 container statuses recorded)&#xA;May 29 00:40:56.577: INFO: &#x9;Container webserver ready: true, restart count 0&#xA;W0529 00:40:56.591085      26 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled.&#xA;May 29 00:40:56.620: INFO: &#xA;Latency metrics for node node2&#xA;May 29 00:40:56.621: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready&#xA;STEP: Destroying namespace &#34;nettest-6329&#34; for this suite.&#xA;</system-out>
      </testcase>
  </testsuite>